[ 514.670376] env[65385]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=65385) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 514.670708] env[65385]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=65385) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 514.670836] env[65385]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=65385) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 514.671245] env[65385]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 514.788035] env[65385]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=65385) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 514.797691] env[65385]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=65385) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 514.843438] env[65385]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 515.401933] env[65385]: INFO nova.virt.driver [None req-f97c78eb-3225-451c-9662-69662b2c13fc None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 515.474816] env[65385]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 515.475033] env[65385]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 515.475138] env[65385]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=65385) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 518.666481] env[65385]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-9521ce7b-6fe2-4394-a192-4d35752dc72b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.683561] env[65385]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=65385) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 518.685088] env[65385]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-f03cb237-72e5-4668-b566-485b27770326 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.718409] env[65385]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 57c11. [ 518.718802] env[65385]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.244s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 518.719495] env[65385]: INFO nova.virt.vmwareapi.driver [None req-f97c78eb-3225-451c-9662-69662b2c13fc None None] VMware vCenter version: 7.0.3 [ 518.723287] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc10c34-1344-4aa2-bd68-d5f0c99983b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.742087] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e283bdd0-a260-48a8-adcf-6122f4f62239 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.749099] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3000c130-944f-49fb-8d5e-c74829689b97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.756919] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a24f31-b729-413e-9c56-ad603a3250a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.770762] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bb439f-2f59-4690-bafe-df8f59a15141 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.778411] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881a8263-aab6-4eaa-b63b-77961f7efc88 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.809688] env[65385]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-f746938c-7448-4dc3-85f9-e304ba742e1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.816738] env[65385]: DEBUG nova.virt.vmwareapi.driver [None req-f97c78eb-3225-451c-9662-69662b2c13fc None None] Extension org.openstack.compute already exists. {{(pid=65385) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 518.820806] env[65385]: INFO nova.compute.provider_config [None req-f97c78eb-3225-451c-9662-69662b2c13fc None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 519.324174] env[65385]: DEBUG nova.context [None req-f97c78eb-3225-451c-9662-69662b2c13fc None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),4d7edf91-ed56-4623-8702-93f73574b3e6(cell1) {{(pid=65385) load_cells /opt/stack/nova/nova/context.py:472}} [ 519.324174] env[65385]: INFO nova.utils [None req-f97c78eb-3225-451c-9662-69662b2c13fc None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 519.325535] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 519.325938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 519.326740] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 519.327328] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Acquiring lock "4d7edf91-ed56-4623-8702-93f73574b3e6" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 519.327644] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Lock "4d7edf91-ed56-4623-8702-93f73574b3e6" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 519.328744] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Lock "4d7edf91-ed56-4623-8702-93f73574b3e6" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 519.349950] env[65385]: INFO dbcounter [None req-aa717398-fd34-4976-871c-73668d594431 None None] Registered counter for database nova_cell0 [ 519.358397] env[65385]: INFO dbcounter [None req-aa717398-fd34-4976-871c-73668d594431 None None] Registered counter for database nova_cell1 [ 519.361968] env[65385]: DEBUG oslo_db.sqlalchemy.engines [None req-aa717398-fd34-4976-871c-73668d594431 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65385) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 519.362210] env[65385]: DEBUG oslo_db.sqlalchemy.engines [None req-aa717398-fd34-4976-871c-73668d594431 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65385) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 519.367193] env[65385]: ERROR nova.db.main.api [None req-aa717398-fd34-4976-871c-73668d594431 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 519.367193] env[65385]: func(*args, **kwargs) [ 519.367193] env[65385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 519.367193] env[65385]: self.work.run() [ 519.367193] env[65385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 519.367193] env[65385]: result = self.fn(*self.args, **self.kwargs) [ 519.367193] env[65385]: File "/opt/stack/nova/nova/utils.py", line 584, in context_wrapper [ 519.367193] env[65385]: return func(*args, **kwargs) [ 519.367193] env[65385]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 519.367193] env[65385]: result = fn(*args, **kwargs) [ 519.367193] env[65385]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 519.367193] env[65385]: return f(*args, **kwargs) [ 519.367193] env[65385]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 519.367193] env[65385]: return db.service_get_minimum_version(context, binaries) [ 519.367193] env[65385]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 519.367193] env[65385]: _check_db_access() [ 519.367193] env[65385]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 519.367193] env[65385]: stacktrace = ''.join(traceback.format_stack()) [ 519.367193] env[65385]: [ 519.370550] env[65385]: ERROR nova.db.main.api [None req-aa717398-fd34-4976-871c-73668d594431 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 519.370550] env[65385]: func(*args, **kwargs) [ 519.370550] env[65385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 519.370550] env[65385]: self.work.run() [ 519.370550] env[65385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 519.370550] env[65385]: result = self.fn(*self.args, **self.kwargs) [ 519.370550] env[65385]: File "/opt/stack/nova/nova/utils.py", line 584, in context_wrapper [ 519.370550] env[65385]: return func(*args, **kwargs) [ 519.370550] env[65385]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 519.370550] env[65385]: result = fn(*args, **kwargs) [ 519.370550] env[65385]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 519.370550] env[65385]: return f(*args, **kwargs) [ 519.370550] env[65385]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 519.370550] env[65385]: return db.service_get_minimum_version(context, binaries) [ 519.370550] env[65385]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 519.370550] env[65385]: _check_db_access() [ 519.370550] env[65385]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 519.370550] env[65385]: stacktrace = ''.join(traceback.format_stack()) [ 519.370550] env[65385]: [ 519.370550] env[65385]: WARNING nova.objects.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 519.370550] env[65385]: WARNING nova.objects.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] Failed to get minimum service version for cell 4d7edf91-ed56-4623-8702-93f73574b3e6 [ 519.370550] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Acquiring lock "singleton_lock" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.371131] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Acquired lock "singleton_lock" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 519.371131] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa717398-fd34-4976-871c-73668d594431 None None] Releasing lock "singleton_lock" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 519.371131] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] Full set of CONF: {{(pid=65385) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/service.py:274}} [ 519.371131] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ******************************************************************************** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 519.371242] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] Configuration options gathered from: {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 519.371315] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 519.371611] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 519.371848] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ================================================================================ {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 519.372260] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] allow_resize_to_same_host = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.372564] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] arq_binding_timeout = 300 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.372797] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] backdoor_port = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.373028] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] backdoor_socket = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.373286] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] block_device_allocate_retries = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.373545] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] block_device_allocate_retries_interval = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.373813] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cell_worker_thread_pool_size = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.374095] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cert = self.pem {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.374413] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.374683] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute_monitors = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.374943] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] config_dir = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.375241] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] config_drive_format = iso9660 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.375470] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.375722] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] config_source = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.375997] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] console_host = devstack {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.376294] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] control_exchange = nova {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.376577] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cpu_allocation_ratio = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.376829] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] daemon = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.377107] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] debug = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.377375] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] default_access_ip_network_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.377646] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] default_availability_zone = nova {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.377934] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] default_ephemeral_format = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.378250] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] default_green_pool_size = 1000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.378637] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.378984] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] default_schedule_zone = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.379270] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] default_thread_pool_size = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380016] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] disk_allocation_ratio = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380016] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] enable_new_services = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380016] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] flat_injected = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380016] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] force_config_drive = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380173] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] force_raw_images = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380314] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] graceful_shutdown_timeout = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380606] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] heal_instance_info_cache_interval = -1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.380984] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] host = cpu-1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.381288] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.381578] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] initial_disk_allocation_ratio = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.381950] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] initial_ram_allocation_ratio = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.382351] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383175] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instance_build_timeout = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383175] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instance_delete_interval = 300 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383175] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instance_format = [instance: %(uuid)s] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383175] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instance_name_template = instance-%08x {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383371] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instance_usage_audit = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383477] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instance_usage_audit_period = month {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383630] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383783] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] instances_path = /opt/stack/data/nova/instances {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.383934] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] internal_service_availability_zone = internal {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.384113] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] key = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.384276] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] live_migration_retry_count = 30 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.384436] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_color = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.384589] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_config_append = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.384745] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.384894] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_dir = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.385059] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_file = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.385202] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_options = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.385360] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_rotate_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.385519] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_rotate_interval_type = days {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.385673] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] log_rotation_type = none {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.385792] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.385906] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.386092] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.386253] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.386373] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.386524] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] long_rpc_timeout = 1800 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.386672] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] max_concurrent_builds = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.386828] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] max_concurrent_live_migrations = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.386968] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] max_concurrent_snapshots = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.387131] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] max_local_block_devices = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.387279] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] max_logfile_count = 30 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.387425] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] max_logfile_size_mb = 200 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.387571] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] maximum_instance_delete_attempts = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.387717] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] migrate_max_retries = -1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.387868] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] mkisofs_cmd = genisoimage {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.388074] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] my_block_storage_ip = 10.180.1.21 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.388203] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] my_ip = 10.180.1.21 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.388389] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.388539] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] network_allocate_retries = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.388704] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.388860] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] osapi_compute_unique_server_name_scope = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.389014] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] password_length = 12 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.389168] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] periodic_enable = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.389320] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] periodic_fuzzy_delay = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.389470] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] pointer_model = usbtablet {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.389621] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] preallocate_images = none {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.389769] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] publish_errors = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.389886] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] pybasedir = /opt/stack/nova {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.390051] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ram_allocation_ratio = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.390207] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] rate_limit_burst = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.390443] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] rate_limit_except_level = CRITICAL {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.390685] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] rate_limit_interval = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.390954] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] reboot_timeout = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.391233] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] reclaim_instance_interval = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.391420] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] record = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.391587] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] reimage_timeout_per_gb = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.391744] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] report_interval = 120 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.391894] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] rescue_timeout = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.392056] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] reserved_host_cpus = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.392207] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] reserved_host_disk_mb = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.392378] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] reserved_host_memory_mb = 512 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.392535] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] reserved_huge_pages = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.392682] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] resize_confirm_window = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.392828] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] resize_fs_using_block_device = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.392971] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] resume_guests_state_on_host_boot = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.393192] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.393360] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] rpc_response_timeout = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.393556] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] run_external_periodic_tasks = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.393692] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] running_deleted_instance_action = reap {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.393889] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] running_deleted_instance_poll_interval = 1800 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.394104] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] running_deleted_instance_timeout = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.394282] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler_instance_sync_interval = 120 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.394472] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_down_time = 720 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.394663] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] servicegroup_driver = db {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.394782] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] shell_completion = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.395032] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] shelved_offload_time = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.395270] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] shelved_poll_interval = 3600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.395457] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] shutdown_timeout = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.395611] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] source_is_ipv6 = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.395761] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ssl_only = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.396009] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.396180] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] sync_power_state_interval = 600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.396349] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] sync_power_state_pool_size = 1000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.396513] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] syslog_log_facility = LOG_USER {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.396660] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] tempdir = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.396805] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] thread_pool_statistic_period = -1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.396950] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] timeout_nbd = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.397142] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] transport_url = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.397390] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] update_resources_interval = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.397649] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] use_cow_images = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.397859] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] use_journal = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.398030] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] use_json = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.398216] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] use_rootwrap_daemon = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.398378] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] use_stderr = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.398527] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] use_syslog = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.398671] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vcpu_pin_set = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.398856] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plugging_is_fatal = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.399038] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plugging_timeout = 300 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.399198] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] virt_mkfs = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.399349] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] volume_usage_poll_interval = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.399498] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] watch_log_file = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.399650] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] web = /usr/share/spice-html5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 519.399821] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_concurrency.disable_process_locking = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.400475] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.400664] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_brick.lock_path = /opt/stack/data/os_brick {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.400830] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.400985] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.401162] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.401320] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.401478] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.401637] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.401827] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.402103] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.compute_link_prefix = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.402197] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.402339] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.dhcp_domain = novalocal {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.402493] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.enable_instance_password = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.402638] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.glance_link_prefix = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.402793] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.402951] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.403118] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.instance_list_per_project_cells = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.403275] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.list_records_by_skipping_down_cells = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.403428] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.local_metadata_per_cell = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.403581] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.max_limit = 1000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.403732] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.metadata_cache_expiration = 15 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.403891] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.neutron_default_project_id = default {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.404071] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.response_validation = warn {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.404237] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.use_neutron_default_nets = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.404422] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.404582] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.404735] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.404894] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.405062] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.vendordata_dynamic_targets = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.405240] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.vendordata_jsonfile_path = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.405415] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.405598] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.backend = dogpile.cache.memcached {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.405754] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.backend_argument = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.405903] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.backend_expiration_time = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.406073] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.config_prefix = cache.oslo {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.406236] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.dead_timeout = 60.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.406387] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.debug_cache_backend = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.406537] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.enable_retry_client = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.406752] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.enable_socket_keepalive = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.406960] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.enabled = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.407169] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.enforce_fips_mode = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.407329] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.expiration_time = 600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.407517] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.hashclient_retry_attempts = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.407720] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.hashclient_retry_delay = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.407876] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_dead_retry = 300 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.408149] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_password = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.408359] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.408565] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.408737] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_pool_maxsize = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.408930] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.409152] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_sasl_enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.409338] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.409510] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_socket_timeout = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.409673] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.memcache_username = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.409831] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.proxies = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.409983] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.redis_db = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.410148] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.redis_password = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.410326] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.redis_sentinel_service_name = mymaster {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.410560] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.410813] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.redis_server = localhost:6379 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.411052] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.redis_socket_timeout = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.411279] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.redis_username = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.411519] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.retry_attempts = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.411756] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.retry_delay = 0.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.411954] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.socket_keepalive_count = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.412220] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.socket_keepalive_idle = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.412442] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.socket_keepalive_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.412606] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.tls_allowed_ciphers = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.412757] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.tls_cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.412906] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.tls_certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.413071] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.tls_enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.413226] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cache.tls_keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.413385] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.413556] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.auth_type = password {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.413698] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.413861] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.catalog_info = volumev3::publicURL {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.414027] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.414204] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.414365] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.cross_az_attach = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.414515] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.debug = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.414663] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.endpoint_template = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.414813] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.http_retries = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.414962] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.415124] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.415337] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.os_region_name = RegionOne {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.415540] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.415709] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cinder.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.415902] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.416120] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.cpu_dedicated_set = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.416290] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.cpu_shared_set = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.416467] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.image_type_exclude_list = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.416625] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.416777] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.max_concurrent_disk_ops = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.416925] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.max_disk_devices_to_attach = -1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.417086] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.417249] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.417399] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.resource_provider_association_refresh = 300 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.417547] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.417697] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.shutdown_retry_interval = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.417874] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.418089] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] conductor.workers = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.418266] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] console.allowed_origins = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.418418] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] console.ssl_ciphers = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.418580] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] console.ssl_minimum_version = default {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.418736] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] consoleauth.enforce_session_timeout = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.418890] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] consoleauth.token_ttl = 600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.419100] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.419263] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.419414] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.419562] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.419710] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.419917] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.420145] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.420338] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.420502] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.420652] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.420803] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.420947] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.421104] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.421267] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.service_type = accelerator {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.421416] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.421561] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.421704] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.421848] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.422020] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.422173] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] cyborg.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.422334] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.asyncio_connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.422485] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.asyncio_slave_connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.422639] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.backend = sqlalchemy {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.422795] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.422949] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.connection_debug = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.423121] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.connection_parameters = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.423321] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.connection_recycle_time = 3600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.423502] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.connection_trace = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.423658] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.db_inc_retry_interval = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.423807] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.db_max_retries = 20 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.423958] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.db_max_retry_interval = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.424146] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.db_retry_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.424315] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.max_overflow = 50 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.424476] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.max_pool_size = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.424627] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.max_retries = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.424784] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.424929] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.mysql_wsrep_sync_wait = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.425087] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.pool_timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.425237] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.retry_interval = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.425381] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.slave_connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.425527] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.sqlite_synchronous = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.425670] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] database.use_db_reconnect = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.425878] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.asyncio_connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.426110] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.asyncio_slave_connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.426374] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.backend = sqlalchemy {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.426609] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.426861] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.connection_debug = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.427071] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.connection_parameters = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.427242] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.connection_recycle_time = 3600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.427397] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.connection_trace = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.427548] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.db_inc_retry_interval = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.427698] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.db_max_retries = 20 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.427849] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.db_max_retry_interval = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.427997] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.db_retry_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.428162] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.max_overflow = 50 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.428328] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.max_pool_size = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.428494] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.max_retries = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.428677] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.428919] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.429196] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.pool_timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.429460] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.retry_interval = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.429677] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.slave_connection = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.429847] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] api_database.sqlite_synchronous = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.430029] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] devices.enabled_mdev_types = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.430203] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.430367] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ephemeral_storage_encryption.default_format = luks {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.430519] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ephemeral_storage_encryption.enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.430671] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.430829] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.api_servers = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.430981] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.431148] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.431303] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.431450] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.431597] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.431746] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.debug = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.431899] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.default_trusted_certificate_ids = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.432090] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.enable_certificate_validation = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.432212] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.enable_rbd_download = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.432387] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.432579] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.432808] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.433032] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.433247] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.433415] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.num_retries = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.433578] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.rbd_ceph_conf = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.433733] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.rbd_connect_timeout = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.433888] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.rbd_pool = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.434068] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.rbd_user = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.434243] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.434400] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.434550] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.434744] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.service_type = image {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.434935] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.435136] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.435307] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.435454] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.435619] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.435770] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.verify_glance_signatures = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.435917] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] glance.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.436137] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] guestfs.debug = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.436322] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.436513] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.auth_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.436680] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.436862] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.437032] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.437186] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.437335] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.437524] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.437707] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.437858] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.438013] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.438170] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.438317] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.438462] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.438604] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.438758] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.service_type = shared-file-system {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.438936] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.share_apply_policy_timeout = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.439114] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.439265] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.439411] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.439557] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.439725] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.439894] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] manila.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.440145] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] mks.enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.440534] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.440720] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] image_cache.manager_interval = 2400 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.440877] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] image_cache.precache_concurrency = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.441045] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] image_cache.remove_unused_base_images = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.441207] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.441365] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.441527] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] image_cache.subdirectory_name = _base {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.441687] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.api_max_retries = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.441838] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.api_retry_interval = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.441983] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.442147] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.auth_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.442294] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.442438] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.442589] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.442738] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.conductor_group = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.442884] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.443037] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.443188] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.443339] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.443484] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.443629] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.443795] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.443921] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.peer_list = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.444103] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.444254] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.444427] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.serial_console_state_timeout = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.444576] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.444734] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.service_type = baremetal {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.444881] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.shard = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.445043] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.445210] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.445409] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.445567] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.445736] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.445884] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ironic.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.446072] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.446240] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] key_manager.fixed_key = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.446410] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.446562] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.barbican_api_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.446708] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.barbican_endpoint = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.446865] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.barbican_endpoint_type = public {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.447022] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.barbican_region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.447175] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.447397] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.447666] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.447891] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.448071] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.448239] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.number_of_retries = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.448427] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.retry_delay = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.448586] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.send_service_user_token = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.448736] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.448881] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.449041] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.verify_ssl = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.449193] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican.verify_ssl_path = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.449347] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.449497] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.auth_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.449643] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.449788] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.449938] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.450096] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.450245] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.450393] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.450538] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] barbican_service_user.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.450694] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.approle_role_id = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.450842] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.approle_secret_id = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.450999] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.kv_mountpoint = secret {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.451163] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.kv_path = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.451316] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.kv_version = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.451465] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.namespace = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.451611] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.root_token_id = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.451755] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.ssl_ca_crt_file = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.451910] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.timeout = 60.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.452070] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.use_ssl = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.452230] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.452427] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.452592] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.452746] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.452895] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.453054] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.453208] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.453360] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.453505] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.453651] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.453796] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.453948] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.454180] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.454329] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.454490] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.service_type = identity {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.454642] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.454792] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.454942] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.455104] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.455271] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.455418] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] keystone.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.455593] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.ceph_mount_options = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.455891] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.456073] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.connection_uri = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.456230] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.cpu_mode = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.456447] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.cpu_model_extra_flags = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.456638] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.cpu_models = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.456835] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.cpu_power_governor_high = performance {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.457042] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.cpu_power_governor_low = powersave {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.457215] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.cpu_power_management = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.457428] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.457677] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.device_detach_attempts = 8 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.457908] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.device_detach_timeout = 20 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.458093] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.disk_cachemodes = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.458253] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.disk_prefix = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.458410] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.enabled_perf_events = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.458659] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.file_backed_memory = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.458870] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.gid_maps = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.459109] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.hw_disk_discard = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.459286] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.hw_machine_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.459448] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.images_rbd_ceph_conf = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.459602] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.459753] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.459907] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.images_rbd_glance_store_name = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.460076] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.images_rbd_pool = rbd {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.460240] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.images_type = default {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.460412] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.images_volume_group = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.460571] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.inject_key = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.460722] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.inject_partition = -2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.460869] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.inject_password = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.461030] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.iscsi_iface = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.461184] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.iser_use_multipath = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.461386] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_bandwidth = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.461627] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.461824] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_downtime = 500 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.462089] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.462286] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.462540] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_inbound_addr = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.462802] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_parallel_connections = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464152] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464152] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_permit_post_copy = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464152] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_scheme = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464152] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_timeout_action = abort {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464152] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_tunnelled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464152] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_uri = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464403] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.live_migration_with_native_tls = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464403] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.max_queues = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464469] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464669] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.464835] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.nfs_mount_options = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.465148] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.465386] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.465555] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.num_iser_scan_tries = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.465707] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.num_memory_encrypted_guests = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.465859] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.466021] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.num_pcie_ports = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.466183] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.num_volume_scan_tries = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.466338] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.pmem_namespaces = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.466486] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.quobyte_client_cfg = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.466786] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.466955] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rbd_connect_timeout = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.467133] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.467293] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.467442] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rbd_secret_uuid = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.467590] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rbd_user = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.467742] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.467899] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.remote_filesystem_transport = ssh {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.468061] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rescue_image_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.468214] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rescue_kernel_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.468378] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rescue_ramdisk_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.468545] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.468691] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.rx_queue_size = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.468846] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.smbfs_mount_options = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.469141] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.469312] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.snapshot_compression = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.469465] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.snapshot_image_format = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.469676] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.469831] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.sparse_logical_volumes = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.469981] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.swtpm_enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.470150] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.swtpm_group = tss {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.470327] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.swtpm_user = tss {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.470573] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.sysinfo_serial = unique {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.470815] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.tb_cache_size = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.471049] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.tx_queue_size = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.471273] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.uid_maps = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.471502] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.use_virtio_for_bridges = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.471680] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.virt_type = kvm {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.471844] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.volume_clear = zero {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.471999] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.volume_clear_size = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.472168] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.volume_enforce_multipath = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.472343] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.volume_use_multipath = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.472524] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.vzstorage_cache_path = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.472775] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.473041] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.vzstorage_mount_group = qemu {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.473300] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.vzstorage_mount_opts = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.473558] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.473865] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.474061] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.vzstorage_mount_user = stack {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.474250] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.474420] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.474585] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.auth_type = password {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.474731] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.474878] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.475039] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.475218] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.475374] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.475534] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.default_floating_pool = public {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.475684] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.475833] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.extension_sync_interval = 600 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.475979] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.http_retries = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.476143] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.476299] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.476466] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.476626] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.476773] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.476929] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.ovs_bridge = br-int {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.477136] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.physnets = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.477327] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.region_name = RegionOne {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.477481] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.477641] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.service_metadata_proxy = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.477790] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.477944] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.service_type = network {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.478108] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.478261] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.478407] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.478554] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.478766] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.478926] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] neutron.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.479107] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] notifications.bdms_in_notifications = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.479277] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] notifications.default_level = INFO {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.479428] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] notifications.include_share_mapping = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.479588] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] notifications.notification_format = unversioned {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.479746] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] notifications.notify_on_state_change = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.479992] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.480206] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] pci.alias = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.480389] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] pci.device_spec = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.480552] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] pci.report_in_placement = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.480715] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.480875] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.auth_type = password {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.481042] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.481195] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.481344] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.481493] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.481636] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.481782] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.481925] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.default_domain_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.482081] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.default_domain_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.482226] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.domain_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.482369] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.domain_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.482512] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.482660] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.482803] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.482945] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.483102] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.483262] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.password = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.483486] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.project_domain_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.483736] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.project_domain_name = Default {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.483961] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.project_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.484185] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.project_name = service {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.484376] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.region_name = RegionOne {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.484540] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.484692] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.484886] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.service_type = placement {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.485055] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.485206] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.485357] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.485506] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.system_scope = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.485743] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.485946] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.trust_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.486200] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.user_domain_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.486423] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.user_domain_name = Default {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.486673] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.user_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.486950] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.username = nova {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.487208] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.487392] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] placement.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.487571] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.cores = 20 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.487728] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.count_usage_from_placement = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.487927] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.488122] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.injected_file_content_bytes = 10240 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.488289] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.injected_file_path_length = 255 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.488460] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.injected_files = 5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.488619] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.instances = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.488772] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.key_pairs = 100 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.488925] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.metadata_items = 128 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.489131] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.ram = 51200 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.489307] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.recheck_quota = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.489466] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.server_group_members = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.489618] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.server_groups = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.489815] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.490038] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] quota.unified_limits_resource_strategy = require {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.490301] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.490486] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.490642] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.image_metadata_prefilter = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.490792] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.490944] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.max_attempts = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.491111] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.max_placement_results = 1000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.491266] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.491413] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.query_placement_for_image_type_support = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.491562] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.491724] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] scheduler.workers = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.491886] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.492057] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.492230] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.492411] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.492574] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.492726] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.492877] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.493070] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.493258] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.host_subset_size = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.493448] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.493601] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.493774] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.493989] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.494213] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.494374] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.isolated_hosts = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.494528] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.isolated_images = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.494678] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.494894] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.495065] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.495273] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.pci_in_placement = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.495446] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.495602] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.495815] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.496027] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.496196] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.496370] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.496535] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.track_instance_changes = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.496701] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.496860] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] metrics.required = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.497021] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] metrics.weight_multiplier = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.497178] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.497356] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] metrics.weight_setting = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.497802] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.498026] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] serial_console.enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.498208] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] serial_console.port_range = 10000:20000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.498391] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.498573] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.498773] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] serial_console.serialproxy_port = 6083 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.498937] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.499116] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.auth_type = password {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.499269] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.499419] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.499574] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.499725] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.499872] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.500043] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.send_service_user_token = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.500199] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.500367] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] service_user.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.500539] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.agent_enabled = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.500693] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.500995] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.501212] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.501375] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.html5proxy_port = 6082 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.501527] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.image_compression = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.501674] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.jpeg_compression = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.501849] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.playback_compression = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.502041] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.require_secure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.502209] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.server_listen = 127.0.0.1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.502373] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.502635] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.502795] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.streaming_mode = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.502947] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] spice.zlib_compression = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.503128] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] upgrade_levels.baseapi = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.503277] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] upgrade_levels.compute = auto {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.503429] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] upgrade_levels.conductor = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.503574] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] upgrade_levels.scheduler = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.503726] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.503875] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.auth_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.504039] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.504243] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.504432] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.504588] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.504736] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.504884] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.505047] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vendordata_dynamic_auth.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.505213] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.api_retry_count = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.505361] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.ca_file = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.505522] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.cache_prefix = devstack-image-cache {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.505675] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.cluster_name = testcl1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.505828] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.connection_pool_size = 10 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.505976] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.console_delay_seconds = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.506148] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.datastore_regex = ^datastore.* {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.506349] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.506512] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.host_password = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.506669] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.host_port = 443 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.506916] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.host_username = administrator@vsphere.local {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.507123] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.insecure = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.507284] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.integration_bridge = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.507435] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.maximum_objects = 100 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.507620] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.pbm_default_policy = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.507776] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.pbm_enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.507924] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.pbm_wsdl_location = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.508103] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.508257] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.serial_port_proxy_uri = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.508425] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.serial_port_service_uri = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.508588] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.task_poll_interval = 0.5 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.508748] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.use_linked_clone = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.508908] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.vnc_keymap = en-us {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.509076] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.vnc_port = 5900 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.509235] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vmware.vnc_port_total = 10000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.509459] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.auth_schemes = ['none'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.509680] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.509968] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.510159] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.510326] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.novncproxy_port = 6080 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.510504] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.server_listen = 127.0.0.1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.510673] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.510824] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.vencrypt_ca_certs = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.510972] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.vencrypt_client_cert = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.511135] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vnc.vencrypt_client_key = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.511298] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.511447] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.disable_deep_image_inspection = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.511594] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.511741] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.511886] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.512087] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.disable_rootwrap = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.512246] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.enable_numa_live_migration = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.512420] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.512577] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.512726] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.512878] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.libvirt_disable_apic = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.513039] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.513192] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.513341] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.513488] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.513633] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.513778] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.513920] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.514113] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.514256] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.514411] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.514601] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.514746] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] wsgi.secure_proxy_ssl_header = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.514899] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] zvm.ca_file = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.515088] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] zvm.cloud_connector_url = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.515367] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.515535] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] zvm.reachable_timeout = 300 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.515696] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.515861] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.516037] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.connection_string = messaging:// {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.516198] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.enabled = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.516380] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.es_doc_type = notification {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.516543] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.es_scroll_size = 10000 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.516698] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.es_scroll_time = 2m {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.516848] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.filter_error_trace = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.517008] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.hmac_keys = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.517176] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.sentinel_service_name = mymaster {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.517329] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.socket_timeout = 0.1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.517478] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.trace_requests = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.517626] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler.trace_sqlalchemy = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.517795] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler_jaeger.process_tags = {} {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.517942] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler_jaeger.service_name_prefix = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.518134] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] profiler_otlp.service_name_prefix = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.518307] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.518456] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.518603] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.518749] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.518896] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.519055] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.519209] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.519360] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.519506] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.519664] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.519813] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.519969] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.520138] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.520299] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.520466] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.520625] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.520776] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.520927] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.521096] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.521307] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.521548] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.521731] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.521886] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.522074] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.522247] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.522399] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.522546] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.522691] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.522889] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.523073] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.523231] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.ssl = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.523405] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.523610] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.523771] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.523929] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.524122] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.ssl_version = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.524279] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.524458] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.524615] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_notifications.retry = -1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.524778] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.524936] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_messaging_notifications.transport_url = **** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.525117] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.auth_section = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.525284] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.auth_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.525438] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.cafile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.525586] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.certfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.525734] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.collect_timing = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.525878] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.connect_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.526032] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.connect_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.526184] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.endpoint_id = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.526343] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.endpoint_interface = publicURL {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.526489] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.endpoint_override = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.526636] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.endpoint_region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.526782] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.endpoint_service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.526928] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.endpoint_service_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.527089] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.insecure = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.527238] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.keyfile = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.527382] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.max_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.527524] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.min_version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.527666] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.region_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.527809] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.retriable_status_codes = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.527953] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.service_name = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.528110] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.service_type = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.528261] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.split_loggers = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.528405] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.status_code_retries = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.528548] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.status_code_retry_delay = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.528692] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.timeout = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.528836] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.valid_interfaces = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.528981] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_limit.version = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.529150] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_reports.file_event_handler = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.529301] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.529445] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] oslo_reports.log_dir = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.529603] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.529751] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.529895] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.530054] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.530211] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.530363] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.530517] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.530676] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.530822] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_ovs_privileged.group = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.530967] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.531128] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.531285] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.531434] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.531580] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] vif_plug_ovs_privileged.user = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.531739] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.flat_interface = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.531907] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.532089] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.532255] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.532415] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.532571] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.532722] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.532867] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.533039] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.533200] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_ovs.isolate_vif = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.533356] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.533505] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.533659] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.533815] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_ovs.ovsdb_interface = native {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.533962] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] os_vif_ovs.per_port_bridge = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.534155] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] privsep_osbrick.capabilities = [21, 2] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.534314] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] privsep_osbrick.group = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.534462] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] privsep_osbrick.helper_command = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.534611] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] privsep_osbrick.log_daemon_traceback = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.534762] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.534911] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.535078] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] privsep_osbrick.user = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.535263] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.535422] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] nova_sys_admin.group = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.535566] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] nova_sys_admin.helper_command = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.535712] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] nova_sys_admin.log_daemon_traceback = False {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.535862] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.536014] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.536170] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] nova_sys_admin.user = None {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 519.536294] env[65385]: DEBUG oslo_service.backend._eventlet.service [None req-aa717398-fd34-4976-871c-73668d594431 None None] ******************************************************************************** {{(pid=65385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 519.536713] env[65385]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 520.041301] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Getting list of instances from cluster (obj){ [ 520.041301] env[65385]: value = "domain-c8" [ 520.041301] env[65385]: _type = "ClusterComputeResource" [ 520.041301] env[65385]: } {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 520.042774] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59aca21e-7778-4e0e-86b7-5e20738d043d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.052442] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Got total of 0 instances {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 520.053047] env[65385]: WARNING nova.virt.vmwareapi.driver [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 520.053502] env[65385]: INFO nova.virt.node [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Generated node identity 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b [ 520.053724] env[65385]: INFO nova.virt.node [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Wrote node identity 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b to /opt/stack/data/n-cpu-1/compute_id [ 520.557072] env[65385]: WARNING nova.compute.manager [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Compute nodes ['1af23b69-5ce6-4d6c-8591-1b95ecca8a6b'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 521.562381] env[65385]: INFO nova.compute.manager [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 522.568376] env[65385]: WARNING nova.compute.manager [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 522.568733] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 522.568847] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 522.568987] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 522.569151] env[65385]: DEBUG nova.compute.resource_tracker [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 522.570066] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1ba7a4-bee6-42ae-8cb0-75e1d5e10554 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.578903] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a37bbdc-a548-45e6-9f46-7bbb90b75f46 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.595071] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b43af3-6fb2-40aa-8039-b0d44736bb5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.602397] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b219ba82-ef76-4422-b794-453441a8ae05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.631920] env[65385]: DEBUG nova.compute.resource_tracker [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180731MB free_disk=97GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 522.632127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 522.632324] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 523.135226] env[65385]: WARNING nova.compute.resource_tracker [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] No compute node record for cpu-1:1af23b69-5ce6-4d6c-8591-1b95ecca8a6b: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b could not be found. [ 523.639141] env[65385]: INFO nova.compute.resource_tracker [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b [ 525.147672] env[65385]: DEBUG nova.compute.resource_tracker [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 525.148128] env[65385]: DEBUG nova.compute.resource_tracker [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 525.347772] env[65385]: INFO nova.scheduler.client.report [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] [req-e4efc150-5290-4af2-8992-6762fbdf33e3] Created resource provider record via placement API for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 525.367569] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b7bef2-1be5-4c31-908b-29e16b210d8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.376512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5535b88e-c316-4ebf-b415-6d3f08f18f9e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.408909] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcf4bf2-4f80-4cc9-a6cf-08813e6d66eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.417413] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cd8850-e2ca-4d68-a7f0-9ab82b714406 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.431616] env[65385]: DEBUG nova.compute.provider_tree [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 525.975666] env[65385]: DEBUG nova.scheduler.client.report [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 525.975989] env[65385]: DEBUG nova.compute.provider_tree [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 0 to 1 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 525.976214] env[65385]: DEBUG nova.compute.provider_tree [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 526.038639] env[65385]: DEBUG nova.compute.provider_tree [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 1 to 2 during operation: update_traits {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 526.543530] env[65385]: DEBUG nova.compute.resource_tracker [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 526.543968] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.911s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 526.543968] env[65385]: DEBUG nova.service [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Creating RPC server for service compute {{(pid=65385) start /opt/stack/nova/nova/service.py:177}} [ 526.557234] env[65385]: DEBUG nova.service [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] Join ServiceGroup membership for this service compute {{(pid=65385) start /opt/stack/nova/nova/service.py:194}} [ 526.557431] env[65385]: DEBUG nova.servicegroup.drivers.db [None req-63402238-a2aa-4fa1-9df7-a60a2a7cf387 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=65385) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 542.562074] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_power_states {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 543.065604] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Getting list of instances from cluster (obj){ [ 543.065604] env[65385]: value = "domain-c8" [ 543.065604] env[65385]: _type = "ClusterComputeResource" [ 543.065604] env[65385]: } {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 543.067060] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e45213-217f-40c9-8ebd-f6780a79c71e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.076901] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Got total of 0 instances {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 543.077403] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 543.077846] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Getting list of instances from cluster (obj){ [ 543.077846] env[65385]: value = "domain-c8" [ 543.077846] env[65385]: _type = "ClusterComputeResource" [ 543.077846] env[65385]: } {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 543.078798] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee63b4b2-6c97-456b-9c36-4b5f48dc2bb3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.086717] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Got total of 0 instances {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 574.046489] env[65385]: INFO nova.utils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] The default thread pool MainProcess.default is initialized [ 574.047432] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "b72885c3-5146-42a5-82e8-444cfd89413d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 574.048289] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "b72885c3-5146-42a5-82e8-444cfd89413d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 574.550794] env[65385]: DEBUG nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 574.844646] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "504f5660-6715-4c7a-965b-9d2ef1852391" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 574.844918] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "504f5660-6715-4c7a-965b-9d2ef1852391" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 574.853286] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.853572] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.853861] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.853929] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.854111] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.854290] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.854461] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.854627] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 574.854891] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.103019] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 575.103019] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 575.104407] env[65385]: INFO nova.compute.claims [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.349097] env[65385]: DEBUG nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 575.358249] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 575.890146] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 575.967450] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 575.967662] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 575.968958] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 575.969183] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 576.222978] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "065659ae-2163-4d6b-a905-fc99d12b6790" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 576.224800] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "065659ae-2163-4d6b-a905-fc99d12b6790" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 576.279685] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae9f9b9-1fcd-4248-ad09-ff42fc85016e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.292872] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa2538f-9d68-4b80-a43e-ee2d311d2e65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.331631] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7613f9ff-3aac-40a4-bee4-da13a6b9821f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.340276] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7401eb53-522a-4771-9e1d-410a72da112e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.356293] env[65385]: DEBUG nova.compute.provider_tree [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.472110] env[65385]: DEBUG nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 576.482457] env[65385]: DEBUG nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 576.728689] env[65385]: DEBUG nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 576.859879] env[65385]: DEBUG nova.scheduler.client.report [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 577.009725] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 577.009725] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 577.277439] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 577.367373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.264s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 577.371040] env[65385]: DEBUG nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 577.373920] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.016s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 577.374135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 577.374299] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 577.374592] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.485s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 577.376228] env[65385]: INFO nova.compute.claims [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.384574] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578bfb20-27f4-4a0c-ae45-51a00d655ef9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.399371] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b52dc1-fe3c-4f0c-84c7-ed7bdaceaaae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.421215] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8875c9-390f-4c29-97fb-c1aa0a3d2c64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.430010] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de09f00-8dff-4477-8b7f-a176bcffdfee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.466199] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180721MB free_disk=97GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 577.466402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 577.885952] env[65385]: DEBUG nova.compute.utils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 577.887923] env[65385]: DEBUG nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 577.888082] env[65385]: DEBUG nova.network.neutron [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 577.889017] env[65385]: WARNING neutronclient.v2_0.client [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 577.890475] env[65385]: WARNING neutronclient.v2_0.client [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 577.892306] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 577.893076] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 578.404577] env[65385]: DEBUG nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 578.603031] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336ce54e-d377-49f4-9633-6470a174072e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.613663] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3182fd7a-25e0-499a-875c-0cb1943c4621 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.647556] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d936fdac-e32e-4662-8a62-be813b555f54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.657418] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3586f5-663c-4f07-b110-3975cf557e73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.674396] env[65385]: DEBUG nova.compute.provider_tree [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.179449] env[65385]: DEBUG nova.scheduler.client.report [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 579.419511] env[65385]: DEBUG nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 579.459209] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 579.459548] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 579.459713] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 579.459893] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 579.460037] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 579.460176] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 579.460373] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.460517] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 579.460979] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 579.461165] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 579.461326] env[65385]: DEBUG nova.virt.hardware [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 579.462220] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57deb337-bf96-465c-a1c3-ce7f6d617ca7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.472916] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bcfb6a-4a04-4640-ae02-bc79af56649d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.490729] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50ab5d4-281a-4a22-8ede-ead221abd2c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.646093] env[65385]: DEBUG nova.policy [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0286d8467da44042a3adafd787c6f362', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '924c9dd4ac16474886959bac37d2254e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 579.682276] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 579.683077] env[65385]: DEBUG nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 579.685814] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.677s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 579.687483] env[65385]: INFO nova.compute.claims [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.194184] env[65385]: DEBUG nova.compute.utils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 580.197103] env[65385]: DEBUG nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 580.197103] env[65385]: DEBUG nova.network.neutron [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 580.201092] env[65385]: WARNING neutronclient.v2_0.client [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 580.204125] env[65385]: WARNING neutronclient.v2_0.client [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 580.204125] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 580.204125] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 580.698152] env[65385]: DEBUG nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 580.813669] env[65385]: DEBUG nova.network.neutron [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Successfully created port: 2b700413-9737-481b-95bf-92741b9085bd {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 580.836312] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78faab3-346c-4d6a-8e46-95ed72e0b7da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.844944] env[65385]: DEBUG nova.policy [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '464b0f7c8fe3426bacdd58396e49db51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d0195ab78e1455fab0eaf7ea55aea99', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 580.854749] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e88ad7-7fbc-4549-a363-4a9dc8f750f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.890916] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de943756-6b3f-45ae-86e8-c20c34879bcd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.902909] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d25124-fc38-4325-b39d-d6f27138d774 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.926787] env[65385]: DEBUG nova.compute.provider_tree [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.239189] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "42b9edf6-1873-49c3-8074-8eef654ac371" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 581.241575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 581.435581] env[65385]: DEBUG nova.scheduler.client.report [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 581.447859] env[65385]: DEBUG nova.network.neutron [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Successfully created port: 80afc4c9-4d95-4675-a30d-521b87803aa6 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 581.711205] env[65385]: DEBUG nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 581.741019] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 581.741019] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 581.741019] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 581.741270] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 581.741270] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 581.741270] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 581.741270] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.741270] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 581.741446] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 581.741522] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 581.741907] env[65385]: DEBUG nova.virt.hardware [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 581.742396] env[65385]: DEBUG nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 581.745620] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d945e50d-076a-435f-abfc-7c5768844861 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.756499] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643251f1-9723-4c6d-b3e7-f7c1035c7e0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.941768] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 581.942416] env[65385]: DEBUG nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 581.945644] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.936s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 581.946953] env[65385]: INFO nova.compute.claims [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.269333] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 582.447547] env[65385]: DEBUG nova.compute.utils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 582.448982] env[65385]: DEBUG nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 582.449178] env[65385]: DEBUG nova.network.neutron [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 582.449480] env[65385]: WARNING neutronclient.v2_0.client [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 582.449758] env[65385]: WARNING neutronclient.v2_0.client [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 582.450338] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 582.450669] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 582.662807] env[65385]: DEBUG nova.network.neutron [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Successfully updated port: 2b700413-9737-481b-95bf-92741b9085bd {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 582.958290] env[65385]: DEBUG nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 583.063247] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7535d4f-598e-4803-ab93-7fa3836e00d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.071700] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce44000a-5ce2-45f8-abf6-4d55b0cde565 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.105977] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576dfab6-2ae7-4d12-a794-589089d5a285 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.114974] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0194690c-757a-410b-85c5-086b1fd002d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.132966] env[65385]: DEBUG nova.compute.provider_tree [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.165788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "refresh_cache-b72885c3-5146-42a5-82e8-444cfd89413d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.166666] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquired lock "refresh_cache-b72885c3-5146-42a5-82e8-444cfd89413d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 583.166666] env[65385]: DEBUG nova.network.neutron [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 583.220429] env[65385]: DEBUG nova.network.neutron [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Successfully updated port: 80afc4c9-4d95-4675-a30d-521b87803aa6 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 583.466241] env[65385]: DEBUG nova.policy [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '885a7aa885a34e42a9ced014bdf480f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67ed32913e88417ea1b9e543242f988d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 583.636248] env[65385]: DEBUG nova.scheduler.client.report [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 583.670062] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 583.670511] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 583.724422] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "refresh_cache-504f5660-6715-4c7a-965b-9d2ef1852391" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.724528] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired lock "refresh_cache-504f5660-6715-4c7a-965b-9d2ef1852391" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 583.724755] env[65385]: DEBUG nova.network.neutron [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 583.932888] env[65385]: DEBUG nova.network.neutron [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 583.977290] env[65385]: DEBUG nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 584.005855] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 584.007028] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 584.007028] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 584.007028] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 584.007028] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 584.007028] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 584.007249] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.007249] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 584.007249] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 584.007394] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 584.007536] env[65385]: DEBUG nova.virt.hardware [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 584.008449] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc392769-49ba-4735-b2e2-a3b5c436b466 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.018602] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "ef235d8e-6193-4010-ac50-72cd275e965a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 584.018897] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "ef235d8e-6193-4010-ac50-72cd275e965a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 584.020879] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949df8be-40a6-4818-8048-f1046e9a5366 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.039517] env[65385]: DEBUG nova.network.neutron [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Successfully created port: 6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 584.143783] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.198s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 584.144352] env[65385]: DEBUG nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 584.147606] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.870s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 584.148785] env[65385]: INFO nova.compute.claims [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.152242] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 584.152579] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 584.228345] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 584.228968] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 584.527990] env[65385]: DEBUG nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 584.661469] env[65385]: DEBUG nova.compute.utils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 584.665603] env[65385]: DEBUG nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 584.665796] env[65385]: DEBUG nova.network.neutron [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 584.666134] env[65385]: WARNING neutronclient.v2_0.client [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 584.666419] env[65385]: WARNING neutronclient.v2_0.client [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 584.667318] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 584.667410] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 584.754125] env[65385]: DEBUG nova.network.neutron [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 585.058917] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 585.166957] env[65385]: DEBUG nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 585.308879] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2eba650-1fe1-4625-8134-deb0b1040f93 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.318604] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93910e5-f4df-4bdf-8757-ef4b30b31f88 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.355453] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50dff451-7865-408d-923a-5f373895cbc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.365015] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f6f831-bccc-4ae1-9c16-085b09b0feaa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.380439] env[65385]: DEBUG nova.compute.provider_tree [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.694387] env[65385]: DEBUG nova.network.neutron [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Successfully updated port: 6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 585.719715] env[65385]: WARNING neutronclient.v2_0.client [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 585.720662] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 585.722179] env[65385]: WARNING openstack [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 585.741247] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 585.741247] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 585.792992] env[65385]: DEBUG nova.policy [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dacc23181a3b4480a96c0735cb239cf2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e4f3e9e9b50481f992d17be1cd10af7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 585.884374] env[65385]: DEBUG nova.scheduler.client.report [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 586.188603] env[65385]: DEBUG nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 586.198970] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "refresh_cache-6e1eeff2-1587-44c6-a154-f73a2d0f8061" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.198970] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquired lock "refresh_cache-6e1eeff2-1587-44c6-a154-f73a2d0f8061" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 586.198970] env[65385]: DEBUG nova.network.neutron [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 586.224996] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 586.225341] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 586.226420] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 586.226420] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 586.226420] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 586.226420] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 586.226721] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.226942] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 586.227232] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 586.227488] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 586.227762] env[65385]: DEBUG nova.virt.hardware [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 586.229628] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08c8813-8d3b-4c73-80ec-835db0b16543 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.242524] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3083f91-1646-4df1-a762-494ef52f9193 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.264614] env[65385]: DEBUG nova.network.neutron [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Updating instance_info_cache with network_info: [{"id": "2b700413-9737-481b-95bf-92741b9085bd", "address": "fa:16:3e:ac:f6:9e", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b700413-97", "ovs_interfaceid": "2b700413-9737-481b-95bf-92741b9085bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 586.293236] env[65385]: WARNING neutronclient.v2_0.client [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 586.294127] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 586.294537] env[65385]: WARNING openstack [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 586.364235] env[65385]: DEBUG nova.network.neutron [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Successfully created port: 87cd4460-4318-4258-a7b7-e1946f8dc8b8 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 586.393565] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 586.393565] env[65385]: DEBUG nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 586.396244] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.930s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 586.680982] env[65385]: DEBUG nova.network.neutron [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Updating instance_info_cache with network_info: [{"id": "80afc4c9-4d95-4675-a30d-521b87803aa6", "address": "fa:16:3e:e6:3a:32", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.111", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc4c9-4d", "ovs_interfaceid": "80afc4c9-4d95-4675-a30d-521b87803aa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 586.703311] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 586.704697] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 586.769076] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Releasing lock "refresh_cache-b72885c3-5146-42a5-82e8-444cfd89413d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 586.769076] env[65385]: DEBUG nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Instance network_info: |[{"id": "2b700413-9737-481b-95bf-92741b9085bd", "address": "fa:16:3e:ac:f6:9e", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b700413-97", "ovs_interfaceid": "2b700413-9737-481b-95bf-92741b9085bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 586.769515] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:f6:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b700413-9737-481b-95bf-92741b9085bd', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 586.788299] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 586.789232] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a76908a4-2709-4c08-b3e0-d5151194f60a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.793467] env[65385]: DEBUG nova.network.neutron [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 586.805169] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Created folder: OpenStack in parent group-v4. [ 586.805322] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Creating folder: Project (924c9dd4ac16474886959bac37d2254e). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 586.806080] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e4aa67f-a9e7-4ef8-894f-17b56aa19019 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.816148] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Created folder: Project (924c9dd4ac16474886959bac37d2254e) in parent group-v870881. [ 586.816346] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Creating folder: Instances. Parent ref: group-v870882. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 586.816594] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ed1c7a9-ddf1-4147-b9ce-9536cc628a90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.827610] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Created folder: Instances in parent group-v870882. [ 586.827879] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 586.828094] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 586.828310] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23e8e2a2-9d79-4b62-873f-a674cfb50ca2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.850823] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 586.851280] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 586.865574] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 586.865574] env[65385]: value = "task-4453028" [ 586.865574] env[65385]: _type = "Task" [ 586.865574] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.874660] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453028, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.900474] env[65385]: DEBUG nova.compute.utils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 586.905126] env[65385]: DEBUG nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 587.000719] env[65385]: WARNING neutronclient.v2_0.client [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 587.001435] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 587.001780] env[65385]: WARNING openstack [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 587.108675] env[65385]: DEBUG nova.network.neutron [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Updating instance_info_cache with network_info: [{"id": "6cd369b2-83d9-43c0-8ba1-89eb0dd6f802", "address": "fa:16:3e:22:b6:9f", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.189", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cd369b2-83", "ovs_interfaceid": "6cd369b2-83d9-43c0-8ba1-89eb0dd6f802", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 587.184917] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Releasing lock "refresh_cache-504f5660-6715-4c7a-965b-9d2ef1852391" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 587.185385] env[65385]: DEBUG nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Instance network_info: |[{"id": "80afc4c9-4d95-4675-a30d-521b87803aa6", "address": "fa:16:3e:e6:3a:32", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.111", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc4c9-4d", "ovs_interfaceid": "80afc4c9-4d95-4675-a30d-521b87803aa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 587.187456] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:3a:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '80afc4c9-4d95-4675-a30d-521b87803aa6', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 587.194285] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Creating folder: Project (5d0195ab78e1455fab0eaf7ea55aea99). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 587.195195] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a12f3301-f80a-46a2-abea-347c5a359aae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.207612] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Created folder: Project (5d0195ab78e1455fab0eaf7ea55aea99) in parent group-v870881. [ 587.207855] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Creating folder: Instances. Parent ref: group-v870885. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 587.208246] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a3d4b0f-d131-4d7e-9131-f33ed5efe811 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.220247] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Created folder: Instances in parent group-v870885. [ 587.220634] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 587.221071] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 587.221460] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ddfaa2c-abd5-43b3-b10a-3fc106b407f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.257404] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 587.257404] env[65385]: value = "task-4453031" [ 587.257404] env[65385]: _type = "Task" [ 587.257404] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.269490] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453031, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.307963] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "a8c76329-5056-4cf0-ac32-85d46429d3db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 587.308281] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 587.382992] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453028, 'name': CreateVM_Task, 'duration_secs': 0.395138} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.382992] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 587.382992] env[65385]: WARNING neutronclient.v2_0.client [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 587.410053] env[65385]: DEBUG nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 587.437230] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b72885c3-5146-42a5-82e8-444cfd89413d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 587.437394] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 504f5660-6715-4c7a-965b-9d2ef1852391 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 587.437513] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 6e1eeff2-1587-44c6-a154-f73a2d0f8061 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 587.437621] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 587.437738] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 065659ae-2163-4d6b-a905-fc99d12b6790 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 587.530454] env[65385]: DEBUG oslo_vmware.service [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c145418e-b2f2-49d0-b083-14df243c119e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.539909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.540109] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 587.541248] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 587.541381] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c173064d-c057-4e37-aa41-fae127bb0f4f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.549465] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 587.549465] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa7cd8-7617-4bcf-3cf3-f2c80648dcf8" [ 587.549465] env[65385]: _type = "Task" [ 587.549465] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.560278] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa7cd8-7617-4bcf-3cf3-f2c80648dcf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.611971] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Releasing lock "refresh_cache-6e1eeff2-1587-44c6-a154-f73a2d0f8061" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 587.612501] env[65385]: DEBUG nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Instance network_info: |[{"id": "6cd369b2-83d9-43c0-8ba1-89eb0dd6f802", "address": "fa:16:3e:22:b6:9f", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.189", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cd369b2-83", "ovs_interfaceid": "6cd369b2-83d9-43c0-8ba1-89eb0dd6f802", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 587.612903] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:b6:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6cd369b2-83d9-43c0-8ba1-89eb0dd6f802', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 587.623651] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Creating folder: Project (67ed32913e88417ea1b9e543242f988d). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 587.624209] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c97ccbc3-f824-4fec-9394-9e73617a1fca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.635996] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Created folder: Project (67ed32913e88417ea1b9e543242f988d) in parent group-v870881. [ 587.636614] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Creating folder: Instances. Parent ref: group-v870888. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 587.636953] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-539b96fd-3b44-4d42-89bb-23c85542a728 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.649723] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Created folder: Instances in parent group-v870888. [ 587.649723] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 587.649723] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 587.649723] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d12aaf79-8697-4874-95e9-4fada925f25f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.675691] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 587.675691] env[65385]: value = "task-4453034" [ 587.675691] env[65385]: _type = "Task" [ 587.675691] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.684995] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453034, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.771798] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453031, 'name': CreateVM_Task, 'duration_secs': 0.400642} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.773985] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 587.774831] env[65385]: WARNING neutronclient.v2_0.client [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 587.775291] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.812773] env[65385]: DEBUG nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 587.941492] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 42b9edf6-1873-49c3-8074-8eef654ac371 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 588.060581] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 588.060876] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 588.061565] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.061789] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 588.062765] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 588.062765] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 588.063324] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 588.066040] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f502e517-0100-4d3f-a560-e7e854163d9b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.066040] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-099c91b9-634b-4dcb-b32e-717494c2b28f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.071064] env[65385]: DEBUG nova.network.neutron [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Successfully updated port: 87cd4460-4318-4258-a7b7-e1946f8dc8b8 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 588.089627] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 588.089627] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6cd20-81ec-7e75-36dc-7654c13241d6" [ 588.089627] env[65385]: _type = "Task" [ 588.089627] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.099537] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 588.099733] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 588.102542] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e22c88-8242-4e38-a8e2-ba151c6c5c6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.109383] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6cd20-81ec-7e75-36dc-7654c13241d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.114172] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53f9dd35-8684-44f1-896f-b671b1f4c2a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.120418] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 588.120418] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52920ce9-7bbd-07f7-163d-83102b14f13d" [ 588.120418] env[65385]: _type = "Task" [ 588.120418] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.133608] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52920ce9-7bbd-07f7-163d-83102b14f13d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.187622] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453034, 'name': CreateVM_Task, 'duration_secs': 0.415698} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.187805] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 588.188328] env[65385]: WARNING neutronclient.v2_0.client [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 588.188665] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.339067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.422554] env[65385]: DEBUG nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 588.444988] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance ef235d8e-6193-4010-ac50-72cd275e965a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 588.449148] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 588.449474] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 588.449648] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 588.449824] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 588.449959] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 588.450110] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 588.450312] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.450460] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 588.450609] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 588.450767] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 588.450930] env[65385]: DEBUG nova.virt.hardware [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 588.452034] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49590d3-612d-4492-8a40-3f4936687c2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.461010] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ac37e2-ea4b-4834-8d6b-a3321a557034 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.478378] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 588.484772] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Creating folder: Project (f62a8aa4e6504425a68c32ba1c284fc0). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 588.485449] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cb3f6ca-a414-42dd-a096-6997b9229321 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.499373] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Created folder: Project (f62a8aa4e6504425a68c32ba1c284fc0) in parent group-v870881. [ 588.499572] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Creating folder: Instances. Parent ref: group-v870891. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 588.499834] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c388fd4-f3f8-43bc-8205-ecc99462e9b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.516069] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Created folder: Instances in parent group-v870891. [ 588.516360] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 588.516580] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 588.516840] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b200002-9d7e-4bb3-85fb-4298cf8f4354 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.540690] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 588.540690] env[65385]: value = "task-4453037" [ 588.540690] env[65385]: _type = "Task" [ 588.540690] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.551734] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453037, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.575611] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "refresh_cache-a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.576204] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquired lock "refresh_cache-a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 588.576204] env[65385]: DEBUG nova.network.neutron [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 588.602598] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 588.603477] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 588.603477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.603477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 588.603777] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 588.603917] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52271b6c-f4bd-4e3e-97f3-02a47902d1ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.610794] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 588.610794] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52129981-bf51-4732-93eb-5aa655974f27" [ 588.610794] env[65385]: _type = "Task" [ 588.610794] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.621160] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52129981-bf51-4732-93eb-5aa655974f27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.631642] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Preparing fetch location {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 588.632678] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Creating directory with path [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 588.632678] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a16ee5b-3873-4659-bd6d-beb8c63bc1ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.657277] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Created directory with path [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 588.657495] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Fetch image to [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 588.657660] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Downloading image file data c3e9f50a-a10e-4952-ab03-d986996039fa to [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk on the data store datastore2 {{(pid=65385) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 588.658571] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1d2b72-fce9-47f6-a120-6be4d666f3a0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.668811] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da79ab53-a369-4d9a-a869-837893b9b36a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.682191] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea05c428-27a2-44bc-9246-edbe5f8929c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.720373] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d6d67f-23b5-4ed8-a2b3-32edf8cad287 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.726148] env[65385]: DEBUG nova.compute.manager [req-0461483f-062d-45cd-93be-24bbeb6de206 req-0f3300ae-740d-489e-a9d5-5437a5613421 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Received event network-vif-plugged-2b700413-9737-481b-95bf-92741b9085bd {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 588.726357] env[65385]: DEBUG oslo_concurrency.lockutils [req-0461483f-062d-45cd-93be-24bbeb6de206 req-0f3300ae-740d-489e-a9d5-5437a5613421 service nova] Acquiring lock "b72885c3-5146-42a5-82e8-444cfd89413d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.726580] env[65385]: DEBUG oslo_concurrency.lockutils [req-0461483f-062d-45cd-93be-24bbeb6de206 req-0f3300ae-740d-489e-a9d5-5437a5613421 service nova] Lock "b72885c3-5146-42a5-82e8-444cfd89413d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 588.726789] env[65385]: DEBUG oslo_concurrency.lockutils [req-0461483f-062d-45cd-93be-24bbeb6de206 req-0f3300ae-740d-489e-a9d5-5437a5613421 service nova] Lock "b72885c3-5146-42a5-82e8-444cfd89413d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 588.726852] env[65385]: DEBUG nova.compute.manager [req-0461483f-062d-45cd-93be-24bbeb6de206 req-0f3300ae-740d-489e-a9d5-5437a5613421 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] No waiting events found dispatching network-vif-plugged-2b700413-9737-481b-95bf-92741b9085bd {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 588.727045] env[65385]: WARNING nova.compute.manager [req-0461483f-062d-45cd-93be-24bbeb6de206 req-0f3300ae-740d-489e-a9d5-5437a5613421 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Received unexpected event network-vif-plugged-2b700413-9737-481b-95bf-92741b9085bd for instance with vm_state building and task_state spawning. [ 588.732906] env[65385]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-4f3c3545-9604-43d4-a6da-da76cdcd588b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.768591] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Downloading image file data c3e9f50a-a10e-4952-ab03-d986996039fa to the data store datastore2 {{(pid=65385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 588.846145] env[65385]: DEBUG oslo_vmware.rw_handles [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 588.852326] env[65385]: DEBUG nova.compute.manager [req-351b0534-72ff-471f-a811-d83576e3ce5e req-575b3f63-86e1-4482-8ada-2c744bb67056 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Received event network-vif-plugged-80afc4c9-4d95-4675-a30d-521b87803aa6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 588.852604] env[65385]: DEBUG oslo_concurrency.lockutils [req-351b0534-72ff-471f-a811-d83576e3ce5e req-575b3f63-86e1-4482-8ada-2c744bb67056 service nova] Acquiring lock "504f5660-6715-4c7a-965b-9d2ef1852391-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.853066] env[65385]: DEBUG oslo_concurrency.lockutils [req-351b0534-72ff-471f-a811-d83576e3ce5e req-575b3f63-86e1-4482-8ada-2c744bb67056 service nova] Lock "504f5660-6715-4c7a-965b-9d2ef1852391-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 588.853066] env[65385]: DEBUG oslo_concurrency.lockutils [req-351b0534-72ff-471f-a811-d83576e3ce5e req-575b3f63-86e1-4482-8ada-2c744bb67056 service nova] Lock "504f5660-6715-4c7a-965b-9d2ef1852391-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 588.853175] env[65385]: DEBUG nova.compute.manager [req-351b0534-72ff-471f-a811-d83576e3ce5e req-575b3f63-86e1-4482-8ada-2c744bb67056 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] No waiting events found dispatching network-vif-plugged-80afc4c9-4d95-4675-a30d-521b87803aa6 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 588.853341] env[65385]: WARNING nova.compute.manager [req-351b0534-72ff-471f-a811-d83576e3ce5e req-575b3f63-86e1-4482-8ada-2c744bb67056 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Received unexpected event network-vif-plugged-80afc4c9-4d95-4675-a30d-521b87803aa6 for instance with vm_state building and task_state spawning. [ 588.948306] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance a8c76329-5056-4cf0-ac32-85d46429d3db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 588.948566] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 588.948764] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=100GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '5', 'num_vm_building': '5', 'num_task_spawning': '4', 'num_os_type_None': '5', 'num_proj_924c9dd4ac16474886959bac37d2254e': '1', 'io_workload': '5', 'num_proj_5d0195ab78e1455fab0eaf7ea55aea99': '1', 'num_proj_67ed32913e88417ea1b9e543242f988d': '1', 'num_proj_0e4f3e9e9b50481f992d17be1cd10af7': '1', 'num_task_None': '1', 'num_proj_f62a8aa4e6504425a68c32ba1c284fc0': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 589.062041] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453037, 'name': CreateVM_Task, 'duration_secs': 0.340498} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.062041] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 589.062406] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.082237] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 589.082646] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 589.126800] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 589.126800] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 589.126800] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.129337] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bc1823-3e04-4206-82a6-17a1de9e0249 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.132586] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 589.133438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 589.133438] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e07165de-b02d-4d95-8966-76af5416815b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.139177] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 589.139177] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ec5a1b-fc08-ed44-6151-ee1e22d6eaf0" [ 589.139177] env[65385]: _type = "Task" [ 589.139177] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.149640] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b04750-13e5-48ff-9b50-965776070379 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.192099] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 589.192422] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 589.194904] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.201396] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a740408d-159d-420d-9a0c-18a1fe4b2964 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.214479] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77ede3f-9446-436b-898d-a88faf44147a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.232320] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.344693] env[65385]: DEBUG nova.network.neutron [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 589.421528] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 589.421528] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 589.739972] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 589.750081] env[65385]: DEBUG oslo_vmware.rw_handles [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Completed reading data from the image iterator. {{(pid=65385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 589.750284] env[65385]: DEBUG oslo_vmware.rw_handles [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 589.820021] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Downloaded image file data c3e9f50a-a10e-4952-ab03-d986996039fa to vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk on the data store datastore2 {{(pid=65385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 589.820217] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Caching image {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 589.820456] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Copying Virtual Disk [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk to [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 589.820745] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7aaac9d2-39ba-4b1e-8e9e-c5ebf071d2d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.835034] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 589.835034] env[65385]: value = "task-4453038" [ 589.835034] env[65385]: _type = "Task" [ 589.835034] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.842968] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.972677] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 589.973383] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 590.066211] env[65385]: WARNING neutronclient.v2_0.client [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 590.066901] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 590.067347] env[65385]: WARNING openstack [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 590.254826] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 590.255079] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.859s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 590.257090] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.988s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 590.258855] env[65385]: INFO nova.compute.claims [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.328979] env[65385]: DEBUG nova.network.neutron [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Updating instance_info_cache with network_info: [{"id": "87cd4460-4318-4258-a7b7-e1946f8dc8b8", "address": "fa:16:3e:a6:3b:e2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4460-43", "ovs_interfaceid": "87cd4460-4318-4258-a7b7-e1946f8dc8b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 590.350697] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453038, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.477796] env[65385]: DEBUG nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 590.831200] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Releasing lock "refresh_cache-a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 590.831617] env[65385]: DEBUG nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Instance network_info: |[{"id": "87cd4460-4318-4258-a7b7-e1946f8dc8b8", "address": "fa:16:3e:a6:3b:e2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4460-43", "ovs_interfaceid": "87cd4460-4318-4258-a7b7-e1946f8dc8b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 590.832155] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:3b:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87cd4460-4318-4258-a7b7-e1946f8dc8b8', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 590.840503] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Creating folder: Project (0e4f3e9e9b50481f992d17be1cd10af7). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 590.840800] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ae87641-c19d-4ae5-8dcf-a602517532f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.856316] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453038, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.7157} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.856663] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Copied Virtual Disk [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk to [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 590.856888] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Deleting the datastore file [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 590.857380] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a08961f1-15c7-4ec3-bf47-f16d506e2fae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.865225] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 590.865225] env[65385]: value = "task-4453040" [ 590.865225] env[65385]: _type = "Task" [ 590.865225] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.871217] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Created folder: Project (0e4f3e9e9b50481f992d17be1cd10af7) in parent group-v870881. [ 590.871545] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Creating folder: Instances. Parent ref: group-v870894. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 590.872375] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2098fa8e-162c-4aa3-aa4d-86f07b98a13f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.878317] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.888042] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Created folder: Instances in parent group-v870894. [ 590.888251] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 590.888482] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 590.888656] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-733b3153-fb05-4dcb-96d3-e44055b0325e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.909163] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 590.909163] env[65385]: value = "task-4453042" [ 590.909163] env[65385]: _type = "Task" [ 590.909163] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.922128] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453042, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.000144] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 591.300582] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "8ac44124-d8cf-421c-9ae4-c943df94550a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 591.300839] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 591.377406] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.030491} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.381163] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 591.381163] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Moving file from [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85/c3e9f50a-a10e-4952-ab03-d986996039fa to [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa. {{(pid=65385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 591.382325] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-4442f0a7-9376-4c9e-8224-6c64ccd893f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.392593] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 591.392593] env[65385]: value = "task-4453043" [ 591.392593] env[65385]: _type = "Task" [ 591.392593] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.401985] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453043, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.419628] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453042, 'name': CreateVM_Task, 'duration_secs': 0.418038} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.422332] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 591.423087] env[65385]: WARNING neutronclient.v2_0.client [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 591.423511] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.423706] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 591.424049] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 591.424434] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f5f708-3fe9-41c2-bb9b-72c92f415b1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.430497] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 591.430497] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b61dd-3434-2ac2-5797-d78bc6d678be" [ 591.430497] env[65385]: _type = "Task" [ 591.430497] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.444132] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b61dd-3434-2ac2-5797-d78bc6d678be, 'name': SearchDatastore_Task, 'duration_secs': 0.009324} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.444435] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 591.444649] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 591.444847] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.473450] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a8fc5d-a114-4828-ab64-dd2c87d247e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.481933] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada778d8-6801-4c9e-aab5-72eed8027f80 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.518999] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6bd880-7a63-4783-88a6-db45763c9755 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.529845] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46de3502-0e9e-41fa-ae38-f8ee3237542b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.546025] env[65385]: DEBUG nova.compute.provider_tree [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 591.807506] env[65385]: DEBUG nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 591.904763] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453043, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029838} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.908056] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] File moved {{(pid=65385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 591.908056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Cleaning up location [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 591.908056] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Deleting the datastore file [datastore2] vmware_temp/2ae30346-8615-4a4d-b27f-ad3e80b32d85 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 591.908056] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f21c3987-7237-4120-8b76-22f2223206bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.913095] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 591.913095] env[65385]: value = "task-4453044" [ 591.913095] env[65385]: _type = "Task" [ 591.913095] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.922228] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453044, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.071364] env[65385]: ERROR nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [req-8736a668-cfb0-49c2-bb63-0e922d7bc596] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8736a668-cfb0-49c2-bb63-0e922d7bc596"}]} [ 592.090829] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 592.112743] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 592.113469] env[65385]: DEBUG nova.compute.provider_tree [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 592.128677] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 592.151313] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 592.331931] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.402496] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aedd9d5-922c-4ad8-be70-85d891a53534 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.412182] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d74594-0d10-4bd9-b255-ee351309e258 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.425462] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453044, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03412} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.453608] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 592.454627] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2fcf5e4-5a1d-4850-a413-bf07499ebdd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.457537] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40255b4-d051-4eda-a3c3-d1df66f46eba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.467581] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 592.467581] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52831da0-c567-901d-333f-e370fbb6bd20" [ 592.467581] env[65385]: _type = "Task" [ 592.467581] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.469105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d2a318-1a89-4830-be89-a96df1d94e94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.489524] env[65385]: DEBUG nova.compute.provider_tree [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 592.495878] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52831da0-c567-901d-333f-e370fbb6bd20, 'name': SearchDatastore_Task, 'duration_secs': 0.010835} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.496461] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 592.496775] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b72885c3-5146-42a5-82e8-444cfd89413d/b72885c3-5146-42a5-82e8-444cfd89413d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 592.497105] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 592.497301] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 592.497621] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1612d09a-a8e1-475f-959e-1713d471092f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.500332] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df8be33f-e385-400c-8f88-1357fbddf408 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.509062] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 592.509062] env[65385]: value = "task-4453045" [ 592.509062] env[65385]: _type = "Task" [ 592.509062] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.514053] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 592.514258] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 592.515402] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83a17360-4041-408b-9dfb-c7628e017c22 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.530762] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.533310] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 592.533310] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ddab6a-8282-8ab7-1bc7-3de404c34cda" [ 592.533310] env[65385]: _type = "Task" [ 592.533310] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.542619] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ddab6a-8282-8ab7-1bc7-3de404c34cda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.622798] env[65385]: DEBUG nova.compute.manager [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Received event network-changed-2b700413-9737-481b-95bf-92741b9085bd {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 592.623070] env[65385]: DEBUG nova.compute.manager [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Refreshing instance network info cache due to event network-changed-2b700413-9737-481b-95bf-92741b9085bd. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 592.623245] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Acquiring lock "refresh_cache-b72885c3-5146-42a5-82e8-444cfd89413d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.623381] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Acquired lock "refresh_cache-b72885c3-5146-42a5-82e8-444cfd89413d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 592.623526] env[65385]: DEBUG nova.network.neutron [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Refreshing network info cache for port 2b700413-9737-481b-95bf-92741b9085bd {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 592.665146] env[65385]: DEBUG nova.compute.manager [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Received event network-changed-80afc4c9-4d95-4675-a30d-521b87803aa6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 592.665349] env[65385]: DEBUG nova.compute.manager [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Refreshing instance network info cache due to event network-changed-80afc4c9-4d95-4675-a30d-521b87803aa6. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 592.665546] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Acquiring lock "refresh_cache-504f5660-6715-4c7a-965b-9d2ef1852391" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.665680] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Acquired lock "refresh_cache-504f5660-6715-4c7a-965b-9d2ef1852391" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 592.665870] env[65385]: DEBUG nova.network.neutron [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Refreshing network info cache for port 80afc4c9-4d95-4675-a30d-521b87803aa6 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 593.022910] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453045, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.027587] env[65385]: ERROR nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [req-3d767ba6-60e0-4b1b-b56b-e4677d3c2ee4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3d767ba6-60e0-4b1b-b56b-e4677d3c2ee4"}]} [ 593.049508] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ddab6a-8282-8ab7-1bc7-3de404c34cda, 'name': SearchDatastore_Task, 'duration_secs': 0.022266} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.051821] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 593.053618] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c53f9fc-dd99-4cad-a774-f640c7b9e0aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.060933] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 593.060933] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528a4196-dad8-f52e-4d38-92c4816fa020" [ 593.060933] env[65385]: _type = "Task" [ 593.060933] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.070674] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528a4196-dad8-f52e-4d38-92c4816fa020, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.073659] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 593.074188] env[65385]: DEBUG nova.compute.provider_tree [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 97, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 593.090178] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 593.113183] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 593.128382] env[65385]: WARNING neutronclient.v2_0.client [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 593.128865] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 593.129499] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 593.169897] env[65385]: WARNING neutronclient.v2_0.client [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 593.170843] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 593.171529] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 593.326230] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "c5b85574-df14-4a7a-ada0-c9668bf18a33" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 593.326445] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "c5b85574-df14-4a7a-ada0-c9668bf18a33" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 593.352242] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c6b808-c9dc-4b62-aa44-0bd79f57f5f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.362920] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e74c4f-7db6-46d4-ae79-71156b945c5e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.404210] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55abdd5-9063-46a0-9992-8ab741f0e2b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.413907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887fed2f-b18f-44ca-bab8-10846e4b6080 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.429903] env[65385]: DEBUG nova.compute.provider_tree [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 593.521620] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552107} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.521913] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b72885c3-5146-42a5-82e8-444cfd89413d/b72885c3-5146-42a5-82e8-444cfd89413d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 593.522362] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 593.522639] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4181949f-5653-419f-aa86-f2223fffadd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.531431] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 593.531431] env[65385]: value = "task-4453046" [ 593.531431] env[65385]: _type = "Task" [ 593.531431] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.542620] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.572115] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528a4196-dad8-f52e-4d38-92c4816fa020, 'name': SearchDatastore_Task, 'duration_secs': 0.019722} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.572399] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 593.572888] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 504f5660-6715-4c7a-965b-9d2ef1852391/504f5660-6715-4c7a-965b-9d2ef1852391.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 593.573036] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 593.573155] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 593.573386] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-523f87d3-8b03-473d-b3c3-c9e92b0bad96 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.575723] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c245b8d-70c0-41d3-b5ac-3fd85825645c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.584025] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 593.584025] env[65385]: value = "task-4453047" [ 593.584025] env[65385]: _type = "Task" [ 593.584025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.588590] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 593.588778] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 593.589633] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd03c255-854f-4af8-a5ba-fdcc3c67f3c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.599550] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.599884] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 593.599884] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a3f208-432f-d3f2-7024-b8746c09eab6" [ 593.599884] env[65385]: _type = "Task" [ 593.599884] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.609199] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a3f208-432f-d3f2-7024-b8746c09eab6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.982858] env[65385]: DEBUG nova.scheduler.client.report [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 15 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 593.982858] env[65385]: DEBUG nova.compute.provider_tree [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 15 to 16 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 593.982858] env[65385]: DEBUG nova.compute.provider_tree [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 594.045153] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076704} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.048467] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 594.048467] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4035ee4c-314e-4346-b9c2-fb58e8c59b2e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.079033] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] b72885c3-5146-42a5-82e8-444cfd89413d/b72885c3-5146-42a5-82e8-444cfd89413d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 594.079455] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24598ba4-f767-485a-ac73-488f59e07df7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.109135] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453047, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51562} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.111619] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 504f5660-6715-4c7a-965b-9d2ef1852391/504f5660-6715-4c7a-965b-9d2ef1852391.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 594.111797] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 594.112100] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 594.112100] env[65385]: value = "task-4453048" [ 594.112100] env[65385]: _type = "Task" [ 594.112100] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.112330] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ec6e0c3-baa0-4e84-8408-a554b0c1d51d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.126955] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453048, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.134778] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a3f208-432f-d3f2-7024-b8746c09eab6, 'name': SearchDatastore_Task, 'duration_secs': 0.021365} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.138567] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 594.138567] env[65385]: value = "task-4453049" [ 594.138567] env[65385]: _type = "Task" [ 594.138567] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.139660] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d193770c-4dba-498b-a3a7-6216ddd734bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.150437] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 594.150437] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528deec7-4ff6-095c-051e-a87136702255" [ 594.150437] env[65385]: _type = "Task" [ 594.150437] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.155778] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453049, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.165295] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528deec7-4ff6-095c-051e-a87136702255, 'name': SearchDatastore_Task, 'duration_secs': 0.011106} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.165295] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 594.165295] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 6e1eeff2-1587-44c6-a154-f73a2d0f8061/6e1eeff2-1587-44c6-a154-f73a2d0f8061.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 594.165295] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 594.165708] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 594.165708] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7303c68-e5bc-481e-8abc-e2d27027dad0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.172878] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55ab8d2e-d257-434c-85bf-88a2e0d6f57f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.180720] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 594.180720] env[65385]: value = "task-4453050" [ 594.180720] env[65385]: _type = "Task" [ 594.180720] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.182160] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 594.182753] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 594.186552] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d0002a6-73e2-44cf-9c2c-a9c25a76efbc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.190660] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 594.191074] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.206192] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453050, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.208031] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 594.208489] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.216798] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 594.216798] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52df88a3-8607-e23d-d7aa-7c05867e658b" [ 594.216798] env[65385]: _type = "Task" [ 594.216798] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.227963] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52df88a3-8607-e23d-d7aa-7c05867e658b, 'name': SearchDatastore_Task, 'duration_secs': 0.012601} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.228835] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce46f088-23e0-4fc3-aa19-a1e963c2317a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.235727] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 594.235727] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f74ed9-44a6-963b-19e8-70669d482cc4" [ 594.235727] env[65385]: _type = "Task" [ 594.235727] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.247912] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f74ed9-44a6-963b-19e8-70669d482cc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.442853] env[65385]: WARNING neutronclient.v2_0.client [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 594.445326] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 594.445326] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.481842] env[65385]: WARNING neutronclient.v2_0.client [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 594.482522] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 594.482893] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.493020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.236s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 594.493466] env[65385]: DEBUG nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 594.497319] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.438s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 594.500508] env[65385]: INFO nova.compute.claims [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.628845] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453048, 'name': ReconfigVM_Task, 'duration_secs': 0.350397} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.629098] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Reconfigured VM instance instance-00000001 to attach disk [datastore2] b72885c3-5146-42a5-82e8-444cfd89413d/b72885c3-5146-42a5-82e8-444cfd89413d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 594.629820] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a003e358-0b0e-4c46-ad75-196d4c3a50e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.640132] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 594.640132] env[65385]: value = "task-4453051" [ 594.640132] env[65385]: _type = "Task" [ 594.640132] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.652550] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453049, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079061} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.656700] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 594.656700] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453051, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.657131] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9ed8d6-fe35-487e-8eb2-bdb3ab3da465 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.684019] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 504f5660-6715-4c7a-965b-9d2ef1852391/504f5660-6715-4c7a-965b-9d2ef1852391.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 594.684980] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-854072a2-f294-4c21-9673-b734808fb9b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.709101] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453050, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478225} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.710413] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 6e1eeff2-1587-44c6-a154-f73a2d0f8061/6e1eeff2-1587-44c6-a154-f73a2d0f8061.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 594.710651] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 594.710960] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 594.710960] env[65385]: value = "task-4453052" [ 594.710960] env[65385]: _type = "Task" [ 594.710960] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.711167] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d119d386-43ec-4a75-b6a8-78f3ee04583d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.721823] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453052, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.723620] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 594.723620] env[65385]: value = "task-4453053" [ 594.723620] env[65385]: _type = "Task" [ 594.723620] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.730620] env[65385]: DEBUG nova.network.neutron [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Updated VIF entry in instance network info cache for port 2b700413-9737-481b-95bf-92741b9085bd. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 594.731076] env[65385]: DEBUG nova.network.neutron [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Updating instance_info_cache with network_info: [{"id": "2b700413-9737-481b-95bf-92741b9085bd", "address": "fa:16:3e:ac:f6:9e", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b700413-97", "ovs_interfaceid": "2b700413-9737-481b-95bf-92741b9085bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 594.736259] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453053, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.746383] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f74ed9-44a6-963b-19e8-70669d482cc4, 'name': SearchDatastore_Task, 'duration_secs': 0.010337} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.747381] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 594.747659] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 594.747949] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 594.748159] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 594.749014] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7afd92c-ecef-4500-a986-4cfd9a09a156 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.751068] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9184e6d0-0f59-4924-a8f2-cecb098eafbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.761766] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 594.761766] env[65385]: value = "task-4453054" [ 594.761766] env[65385]: _type = "Task" [ 594.761766] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.766581] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 594.766818] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 594.767973] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5258a46c-73cb-4f6d-b100-4e6f28858868 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.774112] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453054, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.778290] env[65385]: DEBUG nova.network.neutron [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Updated VIF entry in instance network info cache for port 80afc4c9-4d95-4675-a30d-521b87803aa6. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 594.778658] env[65385]: DEBUG nova.network.neutron [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Updating instance_info_cache with network_info: [{"id": "80afc4c9-4d95-4675-a30d-521b87803aa6", "address": "fa:16:3e:e6:3a:32", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.111", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc4c9-4d", "ovs_interfaceid": "80afc4c9-4d95-4675-a30d-521b87803aa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 594.782323] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 594.782323] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a1decf-db63-a513-bf8b-65002eb6c7fa" [ 594.782323] env[65385]: _type = "Task" [ 594.782323] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.795211] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a1decf-db63-a513-bf8b-65002eb6c7fa, 'name': SearchDatastore_Task, 'duration_secs': 0.01175} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.797890] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcaf6386-7e4f-43e2-a799-09d3e974b6a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.804497] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 594.804497] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524def77-5a00-0353-e5e1-cf9c2bae9378" [ 594.804497] env[65385]: _type = "Task" [ 594.804497] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.816464] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524def77-5a00-0353-e5e1-cf9c2bae9378, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.999697] env[65385]: DEBUG nova.compute.utils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 595.004741] env[65385]: DEBUG nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 595.005023] env[65385]: DEBUG nova.network.neutron [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 595.005754] env[65385]: WARNING neutronclient.v2_0.client [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 595.005754] env[65385]: WARNING neutronclient.v2_0.client [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 595.006406] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 595.006960] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 595.093384] env[65385]: DEBUG nova.policy [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae527c3a64084ef388ff9415f1a6f883', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99173b1f52aa42f9b0db92366db8c411', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 595.155773] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453051, 'name': Rename_Task, 'duration_secs': 0.186615} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.156061] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 595.156504] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46e459fd-d4d3-4571-ac57-b7f35602ec97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.166180] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 595.166180] env[65385]: value = "task-4453055" [ 595.166180] env[65385]: _type = "Task" [ 595.166180] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.177315] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453055, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.228336] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453052, 'name': ReconfigVM_Task, 'duration_secs': 0.332912} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.231970] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 504f5660-6715-4c7a-965b-9d2ef1852391/504f5660-6715-4c7a-965b-9d2ef1852391.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 595.232910] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e41f67a-5d15-47b8-a80e-8d9ebc32162d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.239777] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Releasing lock "refresh_cache-b72885c3-5146-42a5-82e8-444cfd89413d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 595.240038] env[65385]: DEBUG nova.compute.manager [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Received event network-vif-plugged-87cd4460-4318-4258-a7b7-e1946f8dc8b8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 595.240379] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Acquiring lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 595.240472] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 595.240571] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 595.240725] env[65385]: DEBUG nova.compute.manager [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] No waiting events found dispatching network-vif-plugged-87cd4460-4318-4258-a7b7-e1946f8dc8b8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 595.240885] env[65385]: WARNING nova.compute.manager [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Received unexpected event network-vif-plugged-87cd4460-4318-4258-a7b7-e1946f8dc8b8 for instance with vm_state building and task_state spawning. [ 595.241058] env[65385]: DEBUG nova.compute.manager [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Received event network-changed-87cd4460-4318-4258-a7b7-e1946f8dc8b8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 595.241200] env[65385]: DEBUG nova.compute.manager [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Refreshing instance network info cache due to event network-changed-87cd4460-4318-4258-a7b7-e1946f8dc8b8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 595.241370] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Acquiring lock "refresh_cache-a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.243497] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Acquired lock "refresh_cache-a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 595.243497] env[65385]: DEBUG nova.network.neutron [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Refreshing network info cache for port 87cd4460-4318-4258-a7b7-e1946f8dc8b8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 595.243497] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453053, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07485} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.244553] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 595.244657] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 595.244657] env[65385]: value = "task-4453056" [ 595.244657] env[65385]: _type = "Task" [ 595.244657] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.245548] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ae6c88-5c78-469e-b6f0-1dd1a101470b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.257199] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453056, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.277587] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 6e1eeff2-1587-44c6-a154-f73a2d0f8061/6e1eeff2-1587-44c6-a154-f73a2d0f8061.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 595.281770] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bea10313-2a95-4c44-a686-f43c38f96cff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.298117] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Releasing lock "refresh_cache-504f5660-6715-4c7a-965b-9d2ef1852391" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 595.298399] env[65385]: DEBUG nova.compute.manager [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Received event network-vif-plugged-6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 595.298545] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Acquiring lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 595.299463] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 595.299463] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 595.299463] env[65385]: DEBUG nova.compute.manager [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] No waiting events found dispatching network-vif-plugged-6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 595.299463] env[65385]: WARNING nova.compute.manager [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Received unexpected event network-vif-plugged-6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 for instance with vm_state building and task_state spawning. [ 595.299676] env[65385]: DEBUG nova.compute.manager [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Received event network-changed-6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 595.299833] env[65385]: DEBUG nova.compute.manager [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Refreshing instance network info cache due to event network-changed-6cd369b2-83d9-43c0-8ba1-89eb0dd6f802. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 595.300119] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Acquiring lock "refresh_cache-6e1eeff2-1587-44c6-a154-f73a2d0f8061" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.300290] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Acquired lock "refresh_cache-6e1eeff2-1587-44c6-a154-f73a2d0f8061" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 595.300510] env[65385]: DEBUG nova.network.neutron [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Refreshing network info cache for port 6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 595.315132] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453054, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.315541] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 595.315541] env[65385]: value = "task-4453057" [ 595.315541] env[65385]: _type = "Task" [ 595.315541] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.325032] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524def77-5a00-0353-e5e1-cf9c2bae9378, 'name': SearchDatastore_Task, 'duration_secs': 0.01055} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.325747] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 595.326015] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3/a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 595.326328] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05cfb334-6845-4bf7-bda2-ab3e2929ce67 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.331950] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453057, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.338770] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 595.338770] env[65385]: value = "task-4453058" [ 595.338770] env[65385]: _type = "Task" [ 595.338770] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.349035] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.518796] env[65385]: DEBUG nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 595.650767] env[65385]: DEBUG nova.network.neutron [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Successfully created port: 03771ba6-a616-4715-afb9-19306b1b0903 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 595.677679] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453055, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.751044] env[65385]: WARNING neutronclient.v2_0.client [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 595.751741] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 595.752171] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 595.788743] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453054, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.789091] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453056, 'name': Rename_Task, 'duration_secs': 0.182894} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.793119] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 595.793625] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fabdcee9-8918-4f4c-a19a-07e73a074d74 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.803102] env[65385]: WARNING neutronclient.v2_0.client [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 595.804836] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 595.804836] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 595.811785] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 595.811785] env[65385]: value = "task-4453059" [ 595.811785] env[65385]: _type = "Task" [ 595.811785] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.835194] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453059, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.836347] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f596dfa0-043a-4fba-acf9-de774a134ef6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.838571] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453057, 'name': ReconfigVM_Task, 'duration_secs': 0.331689} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.839098] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 6e1eeff2-1587-44c6-a154-f73a2d0f8061/6e1eeff2-1587-44c6-a154-f73a2d0f8061.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 595.839842] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20ccbc6c-e282-465b-be0a-1b63bf2f7d30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.847829] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a39c57-fa4f-4a8a-b397-e66916db6d59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.855073] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.857748] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 595.857748] env[65385]: value = "task-4453060" [ 595.857748] env[65385]: _type = "Task" [ 595.857748] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.890021] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41f216b-8617-4350-b148-ca92910b685c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.895988] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453060, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.903857] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e96257c-54e5-4281-a3db-942f46b7fc9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.932019] env[65385]: DEBUG nova.compute.provider_tree [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.037277] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.037672] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.130262] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.130772] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.177225] env[65385]: DEBUG oslo_vmware.api [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453055, 'name': PowerOnVM_Task, 'duration_secs': 0.528597} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.177285] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 596.177676] env[65385]: INFO nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Took 16.76 seconds to spawn the instance on the hypervisor. [ 596.177909] env[65385]: DEBUG nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 596.178702] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7b2c48-9c36-4d48-94ae-98167e3329ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.287869] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453054, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.471266} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.288194] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 596.288407] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 596.289079] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab9e12ca-b7c0-430e-823c-943a69bd839a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.297249] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 596.297249] env[65385]: value = "task-4453061" [ 596.297249] env[65385]: _type = "Task" [ 596.297249] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.306671] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453061, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.327625] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453059, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.350730] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453058, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.367544] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453060, 'name': Rename_Task, 'duration_secs': 0.158948} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.367847] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 596.368128] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc99d2a1-63b5-4c86-8e49-8012f536d0d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.375587] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 596.375587] env[65385]: value = "task-4453062" [ 596.375587] env[65385]: _type = "Task" [ 596.375587] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.385612] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453062, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.434573] env[65385]: DEBUG nova.scheduler.client.report [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 596.533584] env[65385]: DEBUG nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 596.566168] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 596.566540] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 596.566892] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 596.567284] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 596.567692] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 596.570024] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 596.570024] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.570024] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 596.570024] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 596.570024] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 596.570287] env[65385]: DEBUG nova.virt.hardware [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 596.570474] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868ee6d3-c320-4060-b2b3-748e959dfbed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.583091] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d704051-431c-48cd-9264-f01b08a20023 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.651649] env[65385]: WARNING neutronclient.v2_0.client [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 596.652413] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.652638] env[65385]: WARNING openstack [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.704290] env[65385]: INFO nova.compute.manager [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Took 21.64 seconds to build instance. [ 596.809652] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453061, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102531} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.810676] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 596.812035] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1edec61b-a91a-426f-ae73-81997b25299e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.842894] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 596.844194] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7eea997e-777d-4d8a-94ad-71cddd9e2e12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.867583] env[65385]: DEBUG oslo_vmware.api [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453059, 'name': PowerOnVM_Task, 'duration_secs': 0.893937} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.867583] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 596.867583] env[65385]: INFO nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Took 15.16 seconds to spawn the instance on the hypervisor. [ 596.867583] env[65385]: DEBUG nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 596.869151] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb4a1ee-3eaf-4d15-83ad-62170fd744ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.875151] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453058, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.374766} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.877596] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3/a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 596.877596] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 596.877596] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 596.877596] env[65385]: value = "task-4453063" [ 596.877596] env[65385]: _type = "Task" [ 596.877596] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.877974] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e03ab6b-7376-42d7-802f-243258bb218a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.900540] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453062, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.906182] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 596.906182] env[65385]: value = "task-4453064" [ 596.906182] env[65385]: _type = "Task" [ 596.906182] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.906519] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453063, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.917814] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453064, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.924443] env[65385]: WARNING neutronclient.v2_0.client [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 596.927698] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.927698] env[65385]: WARNING openstack [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.940482] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 596.941135] env[65385]: DEBUG nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 596.945011] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.605s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 596.946078] env[65385]: INFO nova.compute.claims [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.206877] env[65385]: DEBUG oslo_concurrency.lockutils [None req-29e03908-61db-4aa6-89ce-190a4973a01c tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "b72885c3-5146-42a5-82e8-444cfd89413d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.159s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 597.222809] env[65385]: DEBUG nova.network.neutron [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Updated VIF entry in instance network info cache for port 87cd4460-4318-4258-a7b7-e1946f8dc8b8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 597.223311] env[65385]: DEBUG nova.network.neutron [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Updating instance_info_cache with network_info: [{"id": "87cd4460-4318-4258-a7b7-e1946f8dc8b8", "address": "fa:16:3e:a6:3b:e2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87cd4460-43", "ovs_interfaceid": "87cd4460-4318-4258-a7b7-e1946f8dc8b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 597.278885] env[65385]: DEBUG nova.network.neutron [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Successfully updated port: 03771ba6-a616-4715-afb9-19306b1b0903 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 597.392395] env[65385]: DEBUG oslo_vmware.api [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453062, 'name': PowerOnVM_Task, 'duration_secs': 0.569334} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.392665] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 597.392835] env[65385]: INFO nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Took 13.42 seconds to spawn the instance on the hypervisor. [ 597.393188] env[65385]: DEBUG nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 597.394580] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb8b8ac-7c99-4200-a5dc-915a4ad32213 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.399370] env[65385]: DEBUG nova.network.neutron [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Updated VIF entry in instance network info cache for port 6cd369b2-83d9-43c0-8ba1-89eb0dd6f802. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 597.399674] env[65385]: DEBUG nova.network.neutron [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Updating instance_info_cache with network_info: [{"id": "6cd369b2-83d9-43c0-8ba1-89eb0dd6f802", "address": "fa:16:3e:22:b6:9f", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.189", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6cd369b2-83", "ovs_interfaceid": "6cd369b2-83d9-43c0-8ba1-89eb0dd6f802", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 597.409197] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453063, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.424366] env[65385]: INFO nova.compute.manager [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Took 21.56 seconds to build instance. [ 597.432320] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453064, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079388} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.433039] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 597.433532] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66db19ca-6a1a-4996-b50e-95dbc39e8e52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.451896] env[65385]: DEBUG nova.compute.utils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 597.462187] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3/a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 597.465515] env[65385]: DEBUG nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 597.465515] env[65385]: DEBUG nova.network.neutron [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 597.465732] env[65385]: WARNING neutronclient.v2_0.client [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.466228] env[65385]: WARNING neutronclient.v2_0.client [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.466567] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 597.469494] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 597.474119] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-923cced5-d628-4848-91f7-eb13704e6f56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.503095] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 597.503095] env[65385]: value = "task-4453065" [ 597.503095] env[65385]: _type = "Task" [ 597.503095] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.512966] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453065, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.552686] env[65385]: DEBUG nova.policy [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a73969156939405599fd85c6fe810079', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66d654f07ad84cd580f841a7fd31bb9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 597.710594] env[65385]: DEBUG nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 597.726564] env[65385]: DEBUG oslo_concurrency.lockutils [req-ad628f4f-8df8-44ab-9245-21f7afce019f req-652bae99-573d-4581-99ca-2e2c12ceb004 service nova] Releasing lock "refresh_cache-a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 597.784762] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.784762] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 597.784762] env[65385]: DEBUG nova.network.neutron [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 597.896488] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453063, 'name': ReconfigVM_Task, 'duration_secs': 0.710483} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.897730] env[65385]: DEBUG nova.network.neutron [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Successfully created port: a9fd2b9d-1995-43d9-8550-a27e7828eeb6 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 597.900675] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Reconfigured VM instance instance-00000005 to attach disk [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 597.901857] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17a8d20a-5ded-4f00-8ee8-065370ba348d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.911556] env[65385]: DEBUG oslo_concurrency.lockutils [req-016f257e-df03-4088-8a79-7619beb7c1a7 req-7fb073d8-a571-4b9e-861c-50c2cee66692 service nova] Releasing lock "refresh_cache-6e1eeff2-1587-44c6-a154-f73a2d0f8061" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 597.911556] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 597.911556] env[65385]: value = "task-4453066" [ 597.911556] env[65385]: _type = "Task" [ 597.911556] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.921340] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453066, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.927579] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b93437fc-e521-49c7-8b28-6fa2a3660e44 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "504f5660-6715-4c7a-965b-9d2ef1852391" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.082s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 597.937699] env[65385]: INFO nova.compute.manager [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Took 20.96 seconds to build instance. [ 597.965545] env[65385]: DEBUG nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 598.019493] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453065, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.165381] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "649b4e21-fc55-415c-a1f3-ec724397b874" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 598.165607] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 598.249670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 598.260013] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38229cce-6ae2-4245-a241-b3337f3d832a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.268929] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff49922-ee7b-40b7-9a88-2dbdacfe6722 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.304992] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.305193] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.314790] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5620566d-fad1-4b3e-974b-ef6b9409b558 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.323608] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c265c03b-57f7-4e07-b4b3-3b8d3c027d61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.339555] env[65385]: DEBUG nova.compute.provider_tree [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.423734] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453066, 'name': Rename_Task, 'duration_secs': 0.237863} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.423852] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 598.424100] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6839d8c9-ffea-40a6-8fa2-92eaf669a44f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.431047] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 598.431047] env[65385]: value = "task-4453067" [ 598.431047] env[65385]: _type = "Task" [ 598.431047] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.442327] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a6a0b11-e56b-4af3-b150-d9bda969d167 tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.473s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 598.442327] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.516422] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453065, 'name': ReconfigVM_Task, 'duration_secs': 0.747215} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.516746] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Reconfigured VM instance instance-00000004 to attach disk [datastore2] a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3/a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 598.517647] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6af2f6b-0b59-4ea6-a9b7-479b6cb33137 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.526027] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 598.526027] env[65385]: value = "task-4453068" [ 598.526027] env[65385]: _type = "Task" [ 598.526027] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.537550] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453068, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.635856] env[65385]: DEBUG nova.network.neutron [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 598.668166] env[65385]: DEBUG nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 598.827080] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.827547] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.844486] env[65385]: DEBUG nova.scheduler.client.report [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 598.943247] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453067, 'name': PowerOnVM_Task} progress is 64%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.980389] env[65385]: DEBUG nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 599.005952] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 599.006295] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 599.006497] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 599.006719] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 599.006888] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 599.007068] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 599.007322] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.007527] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 599.007685] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 599.007873] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 599.008558] env[65385]: DEBUG nova.virt.hardware [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 599.008994] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a76def0-e21a-4d43-9830-7bee49eb02ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.018273] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436798bb-a6df-4e88-99d5-833b376e03f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.045717] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453068, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.193440] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.352278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 599.352823] env[65385]: DEBUG nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 599.355688] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.356s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.357117] env[65385]: INFO nova.compute.claims [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.443500] env[65385]: DEBUG oslo_vmware.api [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453067, 'name': PowerOnVM_Task, 'duration_secs': 0.841846} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.443836] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 599.444058] env[65385]: INFO nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Took 11.02 seconds to spawn the instance on the hypervisor. [ 599.444253] env[65385]: DEBUG nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 599.445335] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0e1346-4fc5-4a86-b651-fcd4f008fa0d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.536216] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453068, 'name': Rename_Task, 'duration_secs': 0.834804} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.536495] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 599.536744] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ba14f0a-d169-46d9-8b8c-e29ca74d4ab2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.544716] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 599.544716] env[65385]: value = "task-4453069" [ 599.544716] env[65385]: _type = "Task" [ 599.544716] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.554028] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.576391] env[65385]: DEBUG nova.network.neutron [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Successfully updated port: a9fd2b9d-1995-43d9-8550-a27e7828eeb6 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 599.651523] env[65385]: WARNING neutronclient.v2_0.client [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.652255] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.652640] env[65385]: WARNING openstack [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.838916] env[65385]: DEBUG nova.network.neutron [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance_info_cache with network_info: [{"id": "03771ba6-a616-4715-afb9-19306b1b0903", "address": "fa:16:3e:2b:4a:04", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03771ba6-a6", "ovs_interfaceid": "03771ba6-a616-4715-afb9-19306b1b0903", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 599.862281] env[65385]: DEBUG nova.compute.utils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 599.865871] env[65385]: DEBUG nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 599.866320] env[65385]: DEBUG nova.network.neutron [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 599.866595] env[65385]: WARNING neutronclient.v2_0.client [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.866890] env[65385]: WARNING neutronclient.v2_0.client [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.867935] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.871140] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.959398] env[65385]: DEBUG nova.policy [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1a1e4bb5f5a4c869fb41bb398c39350', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a894ba2e7f7c42d091184408d0b84343', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 599.979066] env[65385]: INFO nova.compute.manager [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Took 22.74 seconds to build instance. [ 600.056065] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453069, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.081106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "refresh_cache-ef235d8e-6193-4010-ac50-72cd275e965a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.081306] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquired lock "refresh_cache-ef235d8e-6193-4010-ac50-72cd275e965a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 600.081476] env[65385]: DEBUG nova.network.neutron [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 600.157595] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3de2d1-ffaa-4f53-8d12-c45278b20752 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.170198] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c35521c-8c83-4f0b-9ecb-741e061fa166 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.212040] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da91f7a-27a5-48ca-8f48-f961d362c4c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.221730] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911f37a3-8f8f-4ec7-bafb-baf3e4e14a19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.240428] env[65385]: DEBUG nova.compute.provider_tree [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.342154] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 600.342538] env[65385]: DEBUG nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Instance network_info: |[{"id": "03771ba6-a616-4715-afb9-19306b1b0903", "address": "fa:16:3e:2b:4a:04", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03771ba6-a6", "ovs_interfaceid": "03771ba6-a616-4715-afb9-19306b1b0903", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 600.343032] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:4a:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03771ba6-a616-4715-afb9-19306b1b0903', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 600.351831] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Creating folder: Project (99173b1f52aa42f9b0db92366db8c411). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.351831] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae7cccf3-5b28-4b10-9675-a4ea1dba6392 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.366126] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Created folder: Project (99173b1f52aa42f9b0db92366db8c411) in parent group-v870881. [ 600.366126] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Creating folder: Instances. Parent ref: group-v870897. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.366333] env[65385]: DEBUG nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 600.370579] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f3d35ed-872e-4404-9b86-92863ad719ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.389563] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Created folder: Instances in parent group-v870897. [ 600.389830] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 600.390157] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 600.390728] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10f69cbc-fb88-43c0-96a1-5ec703375ad7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.417785] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 600.417785] env[65385]: value = "task-4453072" [ 600.417785] env[65385]: _type = "Task" [ 600.417785] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.427988] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453072, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.482206] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b797c957-306a-4540-ae72-3f1f53433bc4 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "065659ae-2163-4d6b-a905-fc99d12b6790" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.259s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 600.487559] env[65385]: DEBUG nova.network.neutron [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Successfully created port: 6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 600.555840] env[65385]: DEBUG oslo_vmware.api [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453069, 'name': PowerOnVM_Task, 'duration_secs': 0.844692} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.556143] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 600.556356] env[65385]: INFO nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Took 14.37 seconds to spawn the instance on the hypervisor. [ 600.556539] env[65385]: DEBUG nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 600.557332] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e0e128-7f68-4180-8a24-b4a84cd20abe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.585376] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.585788] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 600.661500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "f77d861a-ae66-44c3-901d-e499926d6f5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.661805] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 600.745595] env[65385]: DEBUG nova.scheduler.client.report [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 600.789949] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "0115dde8-b0aa-4960-a452-ced6587bc567" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.790181] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "0115dde8-b0aa-4960-a452-ced6587bc567" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 600.934930] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453072, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.949465] env[65385]: DEBUG nova.network.neutron [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 601.076891] env[65385]: INFO nova.compute.manager [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Took 24.09 seconds to build instance. [ 601.122835] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.123260] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.165146] env[65385]: DEBUG nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 601.250924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.895s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 601.251491] env[65385]: DEBUG nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 601.254272] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.922s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 601.255806] env[65385]: INFO nova.compute.claims [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.295393] env[65385]: DEBUG nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 601.385802] env[65385]: DEBUG nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 601.416767] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 601.417032] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 601.417214] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 601.417399] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 601.417539] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 601.417679] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 601.417882] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.418046] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 601.418213] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 601.418369] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 601.418532] env[65385]: DEBUG nova.virt.hardware [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 601.419578] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401e3289-3ce7-485d-9bc3-95f111f36881 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.434771] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65694bee-64dd-4b5b-8854-9ee4ccaf6d1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.438805] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453072, 'name': CreateVM_Task, 'duration_secs': 0.673942} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.438961] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 601.440767] env[65385]: WARNING neutronclient.v2_0.client [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.440767] env[65385]: DEBUG oslo_vmware.service [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92c3d8f-c1c8-432c-8276-9f684c450e6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.459346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.459525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.459903] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 601.460184] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85aec47f-9a00-49e4-97a5-cdce4ef1522a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.465610] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 601.465610] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e37e17-0442-9ebf-7e9e-915d889b5ef2" [ 601.465610] env[65385]: _type = "Task" [ 601.465610] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.474819] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e37e17-0442-9ebf-7e9e-915d889b5ef2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.581420] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a6e8de2f-5063-42e9-9eec-bd453faae53b tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.612s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 601.689089] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 601.728549] env[65385]: WARNING neutronclient.v2_0.client [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.729296] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.729698] env[65385]: WARNING openstack [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.762139] env[65385]: DEBUG nova.compute.utils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 601.763066] env[65385]: DEBUG nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 601.763311] env[65385]: DEBUG nova.network.neutron [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 601.763827] env[65385]: WARNING neutronclient.v2_0.client [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.764217] env[65385]: WARNING neutronclient.v2_0.client [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.764990] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.765419] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.821118] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 601.979444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.979953] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 601.980232] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.980410] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.980616] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 601.981410] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-034fb690-cf81-43f1-a9a7-5fda21a3657e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.000472] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 602.000824] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 602.001512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2638c9-ca46-4540-8f74-64a9a2add191 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.013113] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25f17d35-9346-4891-ba1f-1fdcf726148f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.019800] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 602.019800] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e6ce0a-8bea-2f0a-452e-afe1ef64f599" [ 602.019800] env[65385]: _type = "Task" [ 602.019800] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.029890] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e6ce0a-8bea-2f0a-452e-afe1ef64f599, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.110110] env[65385]: DEBUG nova.policy [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c51b0cca7f44f4abcfca5cc312cc7fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73c43e9322314241811db1a37e18f9b4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 602.189993] env[65385]: DEBUG nova.compute.manager [req-d6987acc-3d35-4e53-ba11-3be171c20659 req-90afdba2-5d85-4c3d-99fe-eea03537779f service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Received event network-vif-plugged-03771ba6-a616-4715-afb9-19306b1b0903 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 602.189993] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6987acc-3d35-4e53-ba11-3be171c20659 req-90afdba2-5d85-4c3d-99fe-eea03537779f service nova] Acquiring lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 602.189993] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6987acc-3d35-4e53-ba11-3be171c20659 req-90afdba2-5d85-4c3d-99fe-eea03537779f service nova] Lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 602.189993] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6987acc-3d35-4e53-ba11-3be171c20659 req-90afdba2-5d85-4c3d-99fe-eea03537779f service nova] Lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 602.190566] env[65385]: DEBUG nova.compute.manager [req-d6987acc-3d35-4e53-ba11-3be171c20659 req-90afdba2-5d85-4c3d-99fe-eea03537779f service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] No waiting events found dispatching network-vif-plugged-03771ba6-a616-4715-afb9-19306b1b0903 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 602.190781] env[65385]: WARNING nova.compute.manager [req-d6987acc-3d35-4e53-ba11-3be171c20659 req-90afdba2-5d85-4c3d-99fe-eea03537779f service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Received unexpected event network-vif-plugged-03771ba6-a616-4715-afb9-19306b1b0903 for instance with vm_state building and task_state spawning. [ 602.277597] env[65385]: DEBUG nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 602.291448] env[65385]: DEBUG nova.network.neutron [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Successfully updated port: 6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 602.357617] env[65385]: DEBUG nova.network.neutron [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Updating instance_info_cache with network_info: [{"id": "a9fd2b9d-1995-43d9-8550-a27e7828eeb6", "address": "fa:16:3e:6d:ee:ef", "network": {"id": "c9801743-abb3-4791-8958-57966529d5ac", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-267718441-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66d654f07ad84cd580f841a7fd31bb9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9fd2b9d-19", "ovs_interfaceid": "a9fd2b9d-1995-43d9-8550-a27e7828eeb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 602.539693] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Preparing fetch location {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 602.539957] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Creating directory with path [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 602.540219] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3aa4ecfb-8967-40a2-8c41-ad552bfc34f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.544818] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e5021e-a7b4-4b09-ab68-0570df4d5967 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.552715] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3252c07a-dce1-42d4-a7a6-ed88f3f75401 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.588041] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2870ddfa-267d-459f-82e2-c65a79504314 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.590765] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Created directory with path [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 602.590958] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Fetch image to [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 602.591128] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Downloading image file data c3e9f50a-a10e-4952-ab03-d986996039fa to [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk on the data store datastore1 {{(pid=65385) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 602.591859] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49b6126-9c0f-48d6-b8a2-a8e8582e3309 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.601941] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e69d68f-99ba-406a-9287-ee96695498d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.605585] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b963836-1a8d-4ad9-9616-f8ed47b703c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.625097] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ebabef-0f0b-450e-a59a-cd498f70a7e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.629320] env[65385]: DEBUG nova.compute.provider_tree [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.677469] env[65385]: DEBUG nova.network.neutron [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Successfully created port: 7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 602.682062] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52772dc7-6258-40d1-8376-45682873c8d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.691928] env[65385]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-cb11e3c2-be2d-47ec-9638-af46975bc4bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.727167] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Downloading image file data c3e9f50a-a10e-4952-ab03-d986996039fa to the data store datastore1 {{(pid=65385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 602.787861] env[65385]: DEBUG oslo_vmware.rw_handles [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 602.798307] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.798969] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquired lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 602.798969] env[65385]: DEBUG nova.network.neutron [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 602.879911] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Releasing lock "refresh_cache-ef235d8e-6193-4010-ac50-72cd275e965a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 602.880389] env[65385]: DEBUG nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Instance network_info: |[{"id": "a9fd2b9d-1995-43d9-8550-a27e7828eeb6", "address": "fa:16:3e:6d:ee:ef", "network": {"id": "c9801743-abb3-4791-8958-57966529d5ac", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-267718441-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66d654f07ad84cd580f841a7fd31bb9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9fd2b9d-19", "ovs_interfaceid": "a9fd2b9d-1995-43d9-8550-a27e7828eeb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 602.883734] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.883734] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.896783] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:ee:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8f580e6-1d86-41ee-9ebe-c531cb9299c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a9fd2b9d-1995-43d9-8550-a27e7828eeb6', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 602.910739] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Creating folder: Project (66d654f07ad84cd580f841a7fd31bb9a). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.916582] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53f18b9a-2065-447c-b666-ef3eaf099fa1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.932345] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Created folder: Project (66d654f07ad84cd580f841a7fd31bb9a) in parent group-v870881. [ 602.932530] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Creating folder: Instances. Parent ref: group-v870900. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.932781] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35b76231-07b2-4327-9ffb-806bd3fcae39 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.948633] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Created folder: Instances in parent group-v870900. [ 602.949029] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 602.949548] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 602.949548] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-470dc22e-a8ba-43bc-82c2-0e0464f5b63c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.981736] env[65385]: DEBUG nova.network.neutron [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 602.984553] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 602.984553] env[65385]: value = "task-4453075" [ 602.984553] env[65385]: _type = "Task" [ 602.984553] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.011214] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453075, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.119623] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.120276] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.133795] env[65385]: DEBUG nova.scheduler.client.report [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 603.303028] env[65385]: DEBUG nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 603.345099] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 603.345366] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 603.345544] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 603.345908] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 603.346359] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 603.346359] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 603.346642] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 603.346956] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 603.347116] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 603.347287] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 603.347458] env[65385]: DEBUG nova.virt.hardware [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 603.348520] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a69415-fa16-4c95-91bd-1c13db2cf355 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.356532] env[65385]: WARNING neutronclient.v2_0.client [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.357433] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.357835] env[65385]: WARNING openstack [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.374529] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee398f46-9509-4efb-bdec-c6f0351ddb08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.498648] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453075, 'name': CreateVM_Task, 'duration_secs': 0.415951} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.504367] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 603.508040] env[65385]: WARNING neutronclient.v2_0.client [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.508468] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.508924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 603.509064] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 603.509725] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 603.509980] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 603.511531] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d57cfda9-28e1-4f0d-a51a-399d415f41c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.519335] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 603.519335] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52972e52-d466-172d-140d-24527000e6a0" [ 603.519335] env[65385]: _type = "Task" [ 603.519335] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.537096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 603.537347] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 603.537598] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.634391] env[65385]: DEBUG oslo_vmware.rw_handles [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Completed reading data from the image iterator. {{(pid=65385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 603.634391] env[65385]: DEBUG oslo_vmware.rw_handles [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 603.646025] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 603.646025] env[65385]: DEBUG nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 603.647211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.399s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 603.650189] env[65385]: INFO nova.compute.claims [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.782758] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Downloaded image file data c3e9f50a-a10e-4952-ab03-d986996039fa to vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk on the data store datastore1 {{(pid=65385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 603.784894] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Caching image {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 603.785267] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Copying Virtual Disk [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk to [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 603.785829] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42933ca2-0c5a-460a-a4a1-157e64789ba2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.796543] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 603.796543] env[65385]: value = "task-4453076" [ 603.796543] env[65385]: _type = "Task" [ 603.796543] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.867328] env[65385]: DEBUG nova.network.neutron [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Updating instance_info_cache with network_info: [{"id": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "address": "fa:16:3e:77:44:52", "network": {"id": "8919702d-2cd0-4d13-8f2f-31653142115a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-363641611-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a894ba2e7f7c42d091184408d0b84343", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5cfde5-8f", "ovs_interfaceid": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 604.015045] env[65385]: DEBUG nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 604.104596] env[65385]: DEBUG nova.compute.manager [None req-b19f28aa-39c8-4ec9-8cd2-af72a003094b tempest-ServerDiagnosticsTest-916653588 tempest-ServerDiagnosticsTest-916653588-project-admin] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 604.107555] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823b1e59-04a1-42f2-a979-2adf5fe03f1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.116086] env[65385]: INFO nova.compute.manager [None req-b19f28aa-39c8-4ec9-8cd2-af72a003094b tempest-ServerDiagnosticsTest-916653588 tempest-ServerDiagnosticsTest-916653588-project-admin] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Retrieving diagnostics [ 604.116980] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e7cb03-56a6-4384-8203-f5e7ef3c26e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.167743] env[65385]: DEBUG nova.compute.utils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 604.171507] env[65385]: DEBUG nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 604.171749] env[65385]: DEBUG nova.network.neutron [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 604.172112] env[65385]: WARNING neutronclient.v2_0.client [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.172451] env[65385]: WARNING neutronclient.v2_0.client [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.173169] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 604.173549] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 604.263891] env[65385]: DEBUG nova.policy [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c802fe9c03a34d3a83b4fd98c758d87b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '349d3e2d0d8849819007c19528145c3c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 604.310089] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453076, 'name': CopyVirtualDisk_Task} progress is 27%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.371057] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Releasing lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 604.371588] env[65385]: DEBUG nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Instance network_info: |[{"id": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "address": "fa:16:3e:77:44:52", "network": {"id": "8919702d-2cd0-4d13-8f2f-31653142115a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-363641611-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a894ba2e7f7c42d091184408d0b84343", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5cfde5-8f", "ovs_interfaceid": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 604.372693] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:44:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b5cfde5-8f86-4000-90b9-5c6140ea1606', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 604.383047] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Creating folder: Project (a894ba2e7f7c42d091184408d0b84343). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.384314] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-715d927a-aec6-4292-b8ca-784af5f967db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.397033] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Created folder: Project (a894ba2e7f7c42d091184408d0b84343) in parent group-v870881. [ 604.397589] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Creating folder: Instances. Parent ref: group-v870903. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.397589] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1fe88613-5f39-41ee-a833-06b22abd2dc3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.412743] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Created folder: Instances in parent group-v870903. [ 604.413083] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 604.413994] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 604.413994] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c607a3e-fe07-4acd-bf52-ffbab1180875 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.435683] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 604.435683] env[65385]: value = "task-4453079" [ 604.435683] env[65385]: _type = "Task" [ 604.435683] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.445156] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453079, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.459571] env[65385]: DEBUG nova.network.neutron [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Successfully updated port: 7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 604.541714] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 604.635337] env[65385]: DEBUG nova.network.neutron [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Successfully created port: f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 604.672409] env[65385]: DEBUG nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 604.816406] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453076, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.947215] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453079, 'name': CreateVM_Task, 'duration_secs': 0.421227} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.947442] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 604.948032] env[65385]: WARNING neutronclient.v2_0.client [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.948717] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.948717] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 604.948867] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 604.949746] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7075a68-190c-4da5-9021-0820cc2abfa5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.956823] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 604.956823] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f70ef-daf4-89f2-9e3d-d8ec6c601aa9" [ 604.956823] env[65385]: _type = "Task" [ 604.956823] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.962196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.962196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquired lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 604.962369] env[65385]: DEBUG nova.network.neutron [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 604.968588] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f70ef-daf4-89f2-9e3d-d8ec6c601aa9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.002453] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e4a97e-a396-4b43-afc9-e86fca31b242 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.011772] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cceb14fa-ca06-4516-88ca-5ddb43fa5fc0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.051325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7133bd-c62c-47f1-8819-b769066f314d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.060681] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03513fc2-210b-45d6-b9dc-7d413f2ff259 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.077875] env[65385]: DEBUG nova.compute.provider_tree [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.310153] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453076, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.22066} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.310506] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Copied Virtual Disk [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk to [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 605.310734] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleting the datastore file [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa/tmp-sparse.vmdk {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 605.311102] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-780cd86a-fdd2-4d80-8098-68bec66fc38e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.321018] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 605.321018] env[65385]: value = "task-4453080" [ 605.321018] env[65385]: _type = "Task" [ 605.321018] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.331061] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453080, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.384486] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Acquiring lock "504f5660-6715-4c7a-965b-9d2ef1852391" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.384810] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Lock "504f5660-6715-4c7a-965b-9d2ef1852391" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 605.385055] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Acquiring lock "504f5660-6715-4c7a-965b-9d2ef1852391-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.385835] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Lock "504f5660-6715-4c7a-965b-9d2ef1852391-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 605.385835] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Lock "504f5660-6715-4c7a-965b-9d2ef1852391-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 605.387776] env[65385]: INFO nova.compute.manager [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Terminating instance [ 605.412915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "60067529-1071-4295-b1c9-21523bf347d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.413146] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "60067529-1071-4295-b1c9-21523bf347d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 605.468622] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.468757] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.477384] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 605.477894] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 605.478165] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.580947] env[65385]: DEBUG nova.scheduler.client.report [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 605.679288] env[65385]: INFO nova.compute.manager [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Rebuilding instance [ 605.685651] env[65385]: DEBUG nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 605.713022] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 605.713022] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 605.713022] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 605.713314] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 605.713627] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 605.713972] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 605.714322] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.717027] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 605.717027] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 605.717027] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 605.717027] env[65385]: DEBUG nova.virt.hardware [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 605.717027] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3085d2bf-574c-44a8-8968-92d4ef500364 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.729122] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf75c6d-6cdd-46e3-beb3-73c9dcd58fea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.736956] env[65385]: DEBUG nova.compute.manager [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 605.737907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d7c256-8927-4ce2-b7aa-95457bc02c39 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.831242] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453080, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042629} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.832033] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 605.832372] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Moving file from [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b/c3e9f50a-a10e-4952-ab03-d986996039fa to [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa. {{(pid=65385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 605.832761] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-70b3e179-04fc-416b-8ccf-134b5b189f24 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.840022] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 605.840022] env[65385]: value = "task-4453081" [ 605.840022] env[65385]: _type = "Task" [ 605.840022] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.849681] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453081, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.895285] env[65385]: DEBUG nova.compute.manager [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 605.895285] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.895285] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd61407c-5f21-48c0-9824-89eb5f8d6bdd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.904981] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 605.905520] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d6c9acd-4f92-4dfc-8106-ac59a94d7162 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.913293] env[65385]: DEBUG oslo_vmware.api [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Waiting for the task: (returnval){ [ 605.913293] env[65385]: value = "task-4453082" [ 605.913293] env[65385]: _type = "Task" [ 605.913293] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.922481] env[65385]: DEBUG oslo_vmware.api [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Task: {'id': task-4453082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.089030] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.089030] env[65385]: DEBUG nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 606.090251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.897s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 606.091911] env[65385]: INFO nova.compute.claims [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.272291] env[65385]: DEBUG nova.network.neutron [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Successfully updated port: f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 606.343159] env[65385]: DEBUG nova.network.neutron [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 606.358166] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453081, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.059629} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.358813] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] File moved {{(pid=65385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 606.361022] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Cleaning up location [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 606.361022] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleting the datastore file [datastore1] vmware_temp/95971e53-4a54-4548-9930-3b5d2beaaa1b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 606.361022] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0887b368-7620-4238-b1f4-11809f11ff45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.370440] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 606.370440] env[65385]: value = "task-4453083" [ 606.370440] env[65385]: _type = "Task" [ 606.370440] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.383676] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453083, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.422610] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.425021] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.441789] env[65385]: DEBUG oslo_vmware.api [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Task: {'id': task-4453082, 'name': PowerOffVM_Task, 'duration_secs': 0.504267} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.442428] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 606.442864] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 606.443566] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da15c2f5-c512-45bb-aaa1-0ca5e741d9b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.537563] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 606.537736] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 606.537913] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Deleting the datastore file [datastore2] 504f5660-6715-4c7a-965b-9d2ef1852391 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 606.538215] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9ff47d1a-42c8-4897-be4b-1046e0f34f2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.546418] env[65385]: DEBUG oslo_vmware.api [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Waiting for the task: (returnval){ [ 606.546418] env[65385]: value = "task-4453085" [ 606.546418] env[65385]: _type = "Task" [ 606.546418] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.558073] env[65385]: DEBUG oslo_vmware.api [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Task: {'id': task-4453085, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.597022] env[65385]: DEBUG nova.compute.utils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 606.600612] env[65385]: DEBUG nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 606.718843] env[65385]: DEBUG nova.compute.manager [req-64f466a9-97a4-48b4-87ae-0f6038734f8e req-93de1c51-4298-4d07-bd63-886992ca5962 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Received event network-vif-plugged-6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 606.719097] env[65385]: DEBUG oslo_concurrency.lockutils [req-64f466a9-97a4-48b4-87ae-0f6038734f8e req-93de1c51-4298-4d07-bd63-886992ca5962 service nova] Acquiring lock "a8c76329-5056-4cf0-ac32-85d46429d3db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 606.719464] env[65385]: DEBUG oslo_concurrency.lockutils [req-64f466a9-97a4-48b4-87ae-0f6038734f8e req-93de1c51-4298-4d07-bd63-886992ca5962 service nova] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 606.719671] env[65385]: DEBUG oslo_concurrency.lockutils [req-64f466a9-97a4-48b4-87ae-0f6038734f8e req-93de1c51-4298-4d07-bd63-886992ca5962 service nova] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.719939] env[65385]: DEBUG nova.compute.manager [req-64f466a9-97a4-48b4-87ae-0f6038734f8e req-93de1c51-4298-4d07-bd63-886992ca5962 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] No waiting events found dispatching network-vif-plugged-6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 606.720179] env[65385]: WARNING nova.compute.manager [req-64f466a9-97a4-48b4-87ae-0f6038734f8e req-93de1c51-4298-4d07-bd63-886992ca5962 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Received unexpected event network-vif-plugged-6b5cfde5-8f86-4000-90b9-5c6140ea1606 for instance with vm_state building and task_state spawning. [ 606.761464] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 606.761464] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c9ad66c-679c-45c4-bee4-930d28bf1734 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.768642] env[65385]: WARNING neutronclient.v2_0.client [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 606.769666] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.770105] env[65385]: WARNING openstack [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.781067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.781256] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquired lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.781424] env[65385]: DEBUG nova.network.neutron [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 606.785037] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 606.785037] env[65385]: value = "task-4453086" [ 606.785037] env[65385]: _type = "Task" [ 606.785037] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.797532] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.881703] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453083, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.05129} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.882087] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 606.882853] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d32e3dac-63f3-46e2-a3cc-c22e54b9cdf2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.890169] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 606.890169] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523c9ad0-faa9-f25c-08fb-c597df6bbc82" [ 606.890169] env[65385]: _type = "Task" [ 606.890169] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.900197] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523c9ad0-faa9-f25c-08fb-c597df6bbc82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.947979] env[65385]: DEBUG nova.network.neutron [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updating instance_info_cache with network_info: [{"id": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "address": "fa:16:3e:f9:84:04", "network": {"id": "9ed70dba-8022-422b-8af1-9e125786bbe0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-823001245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73c43e9322314241811db1a37e18f9b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f8bf5-22", "ovs_interfaceid": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 607.062136] env[65385]: DEBUG oslo_vmware.api [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Task: {'id': task-4453085, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268015} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.062136] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 607.062136] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 607.062365] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.063204] env[65385]: INFO nova.compute.manager [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Took 1.17 seconds to destroy the instance on the hypervisor. [ 607.063204] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 607.063204] env[65385]: DEBUG nova.compute.manager [-] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 607.063204] env[65385]: DEBUG nova.network.neutron [-] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 607.064100] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 607.064212] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.064480] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.102590] env[65385]: DEBUG nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 607.284756] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 607.287518] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.288078] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.317039] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453086, 'name': PowerOffVM_Task, 'duration_secs': 0.168218} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.317542] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 607.317542] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.318407] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb0f513-41d0-42f6-9df1-01d43c605ce1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.334467] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 607.334786] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ad4bf9b-63ab-4ae5-affe-c8046a3bf72c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.365222] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 607.365616] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 607.365849] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Deleting the datastore file [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 607.366790] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2917c43c-d805-43b0-8691-7d4d7afeab40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.374163] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 607.374163] env[65385]: value = "task-4453088" [ 607.374163] env[65385]: _type = "Task" [ 607.374163] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.385854] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.401521] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523c9ad0-faa9-f25c-08fb-c597df6bbc82, 'name': SearchDatastore_Task, 'duration_secs': 0.011363} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.401904] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 607.402098] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 42b9edf6-1873-49c3-8074-8eef654ac371/42b9edf6-1873-49c3-8074-8eef654ac371.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 607.402377] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 607.404026] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 607.404026] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4869dfe-4f11-436e-99aa-073f6d372c28 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.406203] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04ec0c73-85c0-47ce-85b7-ade01c03f8b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.414992] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 607.414992] env[65385]: value = "task-4453089" [ 607.414992] env[65385]: _type = "Task" [ 607.414992] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.429142] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 607.429363] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 607.430225] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74d649dd-849b-43f7-bd1c-7649f8e3c47b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.445049] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 607.445049] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52882a72-29bb-8d0d-881e-9cc67cbac39a" [ 607.445049] env[65385]: _type = "Task" [ 607.445049] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.445364] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453089, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.452046] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Releasing lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 607.452497] env[65385]: DEBUG nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Instance network_info: |[{"id": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "address": "fa:16:3e:f9:84:04", "network": {"id": "9ed70dba-8022-422b-8af1-9e125786bbe0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-823001245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73c43e9322314241811db1a37e18f9b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f8bf5-22", "ovs_interfaceid": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 607.452933] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:84:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd88b750a-0e7d-4f16-8bd5-8e6d5743b720', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d3f8bf5-22d2-41b0-9743-0616daaef615', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 607.462175] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Creating folder: Project (73c43e9322314241811db1a37e18f9b4). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 607.463082] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63e7f7b1-b2a7-40db-a2b6-9bdac6cadf46 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.471113] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52882a72-29bb-8d0d-881e-9cc67cbac39a, 'name': SearchDatastore_Task, 'duration_secs': 0.023659} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.472268] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8015e5b6-fffc-4ed6-9301-8567f5dff85c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.477421] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Created folder: Project (73c43e9322314241811db1a37e18f9b4) in parent group-v870881. [ 607.477657] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Creating folder: Instances. Parent ref: group-v870906. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 607.481246] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af54d989-6e73-4a5f-89b0-56b1ad187919 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.485774] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 607.485774] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52638971-a759-8385-323d-5b5e03eef5d8" [ 607.485774] env[65385]: _type = "Task" [ 607.485774] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.497416] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52638971-a759-8385-323d-5b5e03eef5d8, 'name': SearchDatastore_Task, 'duration_secs': 0.012662} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.499566] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 607.499834] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] ef235d8e-6193-4010-ac50-72cd275e965a/ef235d8e-6193-4010-ac50-72cd275e965a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 607.502257] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Created folder: Instances in parent group-v870906. [ 607.502257] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 607.502257] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 607.502257] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 607.502257] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04b85e10-fd71-4aab-9455-cc76559fd431 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.504409] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 607.504409] env[65385]: DEBUG nova.network.neutron [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 607.507180] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-346eb7ed-5c79-498f-b7bf-64b8b57be10c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.508741] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da22fda8-4693-46bc-baa6-b7ded40bf886 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.535543] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 607.535543] env[65385]: value = "task-4453092" [ 607.535543] env[65385]: _type = "Task" [ 607.535543] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.538393] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 607.538600] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 607.539378] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 607.539378] env[65385]: value = "task-4453093" [ 607.539378] env[65385]: _type = "Task" [ 607.539378] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.544680] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0acefc07-3365-4a51-911c-580da4a13460 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.554684] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 607.554684] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52363ff9-a6ab-eadc-fcb0-484222d70251" [ 607.554684] env[65385]: _type = "Task" [ 607.554684] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.563025] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453092, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.563025] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453093, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.575070] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52363ff9-a6ab-eadc-fcb0-484222d70251, 'name': SearchDatastore_Task, 'duration_secs': 0.013405} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.579045] env[65385]: DEBUG nova.compute.manager [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Received event network-changed-03771ba6-a616-4715-afb9-19306b1b0903 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 607.579323] env[65385]: DEBUG nova.compute.manager [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Refreshing instance network info cache due to event network-changed-03771ba6-a616-4715-afb9-19306b1b0903. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 607.579723] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Acquiring lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.579830] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Acquired lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 607.581149] env[65385]: DEBUG nova.network.neutron [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Refreshing network info cache for port 03771ba6-a616-4715-afb9-19306b1b0903 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 607.582735] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4b084a6-2d90-400c-ba61-eaf3dce123bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.586392] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18540d53-c285-4342-a633-250eb601f966 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.595127] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 607.595127] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521ac9bd-71cf-bb72-9290-3b3a3fb14e22" [ 607.595127] env[65385]: _type = "Task" [ 607.595127] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.601709] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eaa42e5-8e32-4f63-8b5e-a64a81577121 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.647105] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521ac9bd-71cf-bb72-9290-3b3a3fb14e22, 'name': SearchDatastore_Task, 'duration_secs': 0.010743} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.648243] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 607.648514] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] a8c76329-5056-4cf0-ac32-85d46429d3db/a8c76329-5056-4cf0-ac32-85d46429d3db.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 607.649841] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fc1c7a-9727-423b-80d2-e07a41e7ddf3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.652524] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4fefef8b-4ab6-4293-9c10-6361f38d4600 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.662463] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc04007-705b-4916-a6b8-b677bdbf0cac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.674921] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 607.674921] env[65385]: value = "task-4453094" [ 607.674921] env[65385]: _type = "Task" [ 607.674921] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.692048] env[65385]: DEBUG nova.compute.provider_tree [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.700750] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.778708] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.778708] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.816310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "59450963-d83f-46e8-8c13-05d4f1818c64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 607.816537] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "59450963-d83f-46e8-8c13-05d4f1818c64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 607.886926] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.116588} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.887319] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 607.887869] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 607.887869] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.930772] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453089, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.055291] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453092, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.062797] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453093, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.090880] env[65385]: WARNING neutronclient.v2_0.client [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 608.092595] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 608.092595] env[65385]: WARNING openstack [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 608.101134] env[65385]: WARNING neutronclient.v2_0.client [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 608.101655] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 608.102091] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 608.150817] env[65385]: DEBUG nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 608.197613] env[65385]: DEBUG nova.scheduler.client.report [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 608.200924] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.203546] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 608.204167] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 608.204167] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 608.204167] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 608.204306] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 608.204457] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 608.204595] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.204801] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 608.204933] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 608.205103] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 608.205263] env[65385]: DEBUG nova.virt.hardware [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 608.206191] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114ac5f0-8058-423e-9c25-6812a3fe2368 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.220375] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5300eb8a-760c-4bc3-8b05-5b8b37cf6c42 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.245056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 608.254154] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Creating folder: Project (05210627ad7d4e2a8250904759479ca7). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 608.255434] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66296824-4f18-4a40-9ab1-f8c7e369ff11 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.271991] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Created folder: Project (05210627ad7d4e2a8250904759479ca7) in parent group-v870881. [ 608.272201] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Creating folder: Instances. Parent ref: group-v870909. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 608.272468] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96383e59-5b92-4de8-84a8-a51803e983ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.285731] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Created folder: Instances in parent group-v870909. [ 608.285971] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 608.290651] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 608.290651] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa229e28-d19d-427f-86a2-cc5dd456a5ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.300087] env[65385]: DEBUG nova.network.neutron [-] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 608.309894] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 608.309894] env[65385]: value = "task-4453097" [ 608.309894] env[65385]: _type = "Task" [ 608.309894] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.322096] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453097, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.434510] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453089, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549346} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.434853] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 42b9edf6-1873-49c3-8074-8eef654ac371/42b9edf6-1873-49c3-8074-8eef654ac371.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 608.435123] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 608.435409] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0626435c-287d-435f-b9d8-5d5eb7470717 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.448651] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 608.448651] env[65385]: value = "task-4453098" [ 608.448651] env[65385]: _type = "Task" [ 608.448651] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.460585] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453098, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.559798] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453093, 'name': CreateVM_Task, 'duration_secs': 0.543836} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.563385] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 608.564110] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453092, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.810538} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.564959] env[65385]: WARNING neutronclient.v2_0.client [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 608.564959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.564959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 608.565760] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 608.565760] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] ef235d8e-6193-4010-ac50-72cd275e965a/ef235d8e-6193-4010-ac50-72cd275e965a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 608.565948] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 608.566212] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-184119b9-d643-473d-9e65-8c16777b020a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.568440] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2eb179ea-6394-4195-9bcf-11623f8ee932 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.577554] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 608.577554] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529a0b51-b134-7e6e-aa30-b7752180919a" [ 608.577554] env[65385]: _type = "Task" [ 608.577554] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.580103] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 608.580103] env[65385]: value = "task-4453099" [ 608.580103] env[65385]: _type = "Task" [ 608.580103] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.597144] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529a0b51-b134-7e6e-aa30-b7752180919a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.601261] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453099, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.696618] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453094, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.011933} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.696618] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] a8c76329-5056-4cf0-ac32-85d46429d3db/a8c76329-5056-4cf0-ac32-85d46429d3db.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 608.696618] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 608.696618] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fcefb27-c6e0-4e6f-98e8-42fab17c474a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.707296] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 608.707296] env[65385]: value = "task-4453100" [ 608.707296] env[65385]: _type = "Task" [ 608.707296] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.712217] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 608.713357] env[65385]: DEBUG nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 608.723446] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.033s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 608.723446] env[65385]: INFO nova.compute.claims [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.732651] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453100, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.803027] env[65385]: INFO nova.compute.manager [-] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Took 1.74 seconds to deallocate network for instance. [ 608.822772] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453097, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.840132] env[65385]: DEBUG nova.network.neutron [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Updating instance_info_cache with network_info: [{"id": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "address": "fa:16:3e:0e:a3:79", "network": {"id": "aec9c0ce-a40e-4d0b-ad1d-2984fe3a4e89", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1834228009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "349d3e2d0d8849819007c19528145c3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5dabab4-c5", "ovs_interfaceid": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 608.911843] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 608.912270] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 608.947313] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 608.947313] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 608.947313] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 608.947313] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 608.947549] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 608.947549] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 608.947779] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.947779] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 608.948258] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 608.948258] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 608.948658] env[65385]: DEBUG nova.virt.hardware [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 608.949631] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410b888e-c5e6-44de-8560-9b20474cc2de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.968234] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453098, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103995} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.969791] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cadddd6d-4e3d-49ef-a0fb-bafa80fb767f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.976025] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 608.985239] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b147dd7-4f92-47f6-bc12-06e6d07768ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.011238] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] 42b9edf6-1873-49c3-8074-8eef654ac371/42b9edf6-1873-49c3-8074-8eef654ac371.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 609.020387] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9855c554-7c4b-4d14-806a-d2346f93fe4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.035835] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 609.041209] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 609.041712] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 609.042312] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b549c18-da3e-4776-b450-4e178db2c73b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.060998] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 609.060998] env[65385]: value = "task-4453101" [ 609.060998] env[65385]: _type = "Task" [ 609.060998] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.067614] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 609.067614] env[65385]: value = "task-4453102" [ 609.067614] env[65385]: _type = "Task" [ 609.067614] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.076496] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453101, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.083110] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453102, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.097702] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529a0b51-b134-7e6e-aa30-b7752180919a, 'name': SearchDatastore_Task, 'duration_secs': 0.05479} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.101037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.101708] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 609.101708] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.101708] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.102035] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 609.102142] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453099, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095222} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.102358] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-840a801b-309c-45c6-82ea-8e85c0663644 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.104483] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 609.105294] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a644d30-0009-4d2b-b68c-5921d9314d43 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.128280] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] ef235d8e-6193-4010-ac50-72cd275e965a/ef235d8e-6193-4010-ac50-72cd275e965a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 609.130099] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aae6bf42-c264-4ee1-9b94-9ea0803223d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.145038] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 609.145038] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 609.145423] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86a23703-fde3-4827-9a27-20eb157a1ef1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.152774] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 609.152774] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c79d3d-844a-d5fd-69db-35eb77f7cf06" [ 609.152774] env[65385]: _type = "Task" [ 609.152774] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.160025] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 609.160025] env[65385]: value = "task-4453103" [ 609.160025] env[65385]: _type = "Task" [ 609.160025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.170725] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c79d3d-844a-d5fd-69db-35eb77f7cf06, 'name': SearchDatastore_Task, 'duration_secs': 0.01157} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.172936] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e341a31b-6793-4a64-b682-b5fda01d6f19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.180496] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453103, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.185798] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 609.185798] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52afbb3a-5191-ca42-d03b-670691938d40" [ 609.185798] env[65385]: _type = "Task" [ 609.185798] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.198933] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52afbb3a-5191-ca42-d03b-670691938d40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.223337] env[65385]: DEBUG nova.compute.utils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 609.225049] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453100, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.141681} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.226408] env[65385]: WARNING neutronclient.v2_0.client [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.227232] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.228215] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.241566] env[65385]: DEBUG nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 609.241566] env[65385]: DEBUG nova.network.neutron [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 609.241942] env[65385]: WARNING neutronclient.v2_0.client [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.242460] env[65385]: WARNING neutronclient.v2_0.client [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.243671] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.244296] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.252251] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 609.256263] env[65385]: DEBUG nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 609.260184] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57cb06c1-c43c-4e2d-9ccc-04c2f65a5a68 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.293548] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] a8c76329-5056-4cf0-ac32-85d46429d3db/a8c76329-5056-4cf0-ac32-85d46429d3db.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 609.296202] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62930a1d-8c4d-4b18-a830-e34cb554871c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.313590] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.326565] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453097, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.335874] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 609.335874] env[65385]: value = "task-4453104" [ 609.335874] env[65385]: _type = "Task" [ 609.335874] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.347022] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Releasing lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.347022] env[65385]: DEBUG nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Instance network_info: |[{"id": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "address": "fa:16:3e:0e:a3:79", "network": {"id": "aec9c0ce-a40e-4d0b-ad1d-2984fe3a4e89", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1834228009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "349d3e2d0d8849819007c19528145c3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5dabab4-c5", "ovs_interfaceid": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 609.347255] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453104, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.348394] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:a3:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd0eb882-ab95-4373-aa20-ee565a9919e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5dabab4-c5af-42f4-bfdc-eec4330bdb0d', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 609.356402] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Creating folder: Project (349d3e2d0d8849819007c19528145c3c). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 609.357111] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a51fe80-112c-42e3-ab6d-e5ba4176cea3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.374023] env[65385]: DEBUG nova.policy [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53765f0467d14fc3bc946cc893f75280', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '112402342b0e481f83be7e17c9f739e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 609.377712] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Created folder: Project (349d3e2d0d8849819007c19528145c3c) in parent group-v870881. [ 609.377908] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Creating folder: Instances. Parent ref: group-v870913. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 609.378177] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d92b0859-1ea8-4114-b30e-85be9aa2a51a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.390222] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Created folder: Instances in parent group-v870913. [ 609.390401] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 609.391248] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 609.391248] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf0d86c1-fc6b-4845-8154-17d411f43a66 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.414933] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 609.414933] env[65385]: value = "task-4453107" [ 609.414933] env[65385]: _type = "Task" [ 609.414933] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.422805] env[65385]: DEBUG nova.network.neutron [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updated VIF entry in instance network info cache for port 03771ba6-a616-4715-afb9-19306b1b0903. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 609.423159] env[65385]: DEBUG nova.network.neutron [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance_info_cache with network_info: [{"id": "03771ba6-a616-4715-afb9-19306b1b0903", "address": "fa:16:3e:2b:4a:04", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03771ba6-a6", "ovs_interfaceid": "03771ba6-a616-4715-afb9-19306b1b0903", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 609.428689] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453107, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.431597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "b72885c3-5146-42a5-82e8-444cfd89413d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.431597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "b72885c3-5146-42a5-82e8-444cfd89413d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.431597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "b72885c3-5146-42a5-82e8-444cfd89413d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.431597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "b72885c3-5146-42a5-82e8-444cfd89413d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.433168] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "b72885c3-5146-42a5-82e8-444cfd89413d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 609.434101] env[65385]: INFO nova.compute.manager [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Terminating instance [ 609.508646] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.508909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.577612] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453101, 'name': ReconfigVM_Task, 'duration_secs': 0.337646} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.577612] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Reconfigured VM instance instance-00000006 to attach disk [datastore1] 42b9edf6-1873-49c3-8074-8eef654ac371/42b9edf6-1873-49c3-8074-8eef654ac371.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 609.577612] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fdaa18c0-0ab6-43aa-800b-4c5bf2b05b84 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.584249] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453102, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.591076] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 609.591076] env[65385]: value = "task-4453108" [ 609.591076] env[65385]: _type = "Task" [ 609.591076] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.603400] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453108, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.674742] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453103, 'name': ReconfigVM_Task, 'duration_secs': 0.344222} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.675119] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Reconfigured VM instance instance-00000007 to attach disk [datastore1] ef235d8e-6193-4010-ac50-72cd275e965a/ef235d8e-6193-4010-ac50-72cd275e965a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 609.676581] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0216ea2-177c-43c1-a122-83941fd0f7ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.685453] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 609.685453] env[65385]: value = "task-4453109" [ 609.685453] env[65385]: _type = "Task" [ 609.685453] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.698698] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453109, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.702583] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52afbb3a-5191-ca42-d03b-670691938d40, 'name': SearchDatastore_Task, 'duration_secs': 0.011699} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.702849] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.703141] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] bf79fbcd-e943-4ff3-bcf8-6a2817df8292/bf79fbcd-e943-4ff3-bcf8-6a2817df8292.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 609.703419] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7cce6b8-3d16-47ce-919f-2aa5cb41926f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.713474] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 609.713474] env[65385]: value = "task-4453110" [ 609.713474] env[65385]: _type = "Task" [ 609.713474] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.725541] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453110, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.828349] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453097, 'name': CreateVM_Task, 'duration_secs': 1.462695} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.828539] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 609.829234] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.829234] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.829402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 609.829662] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7a5c46c-6ab7-4798-a13d-f687f02b571a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.844294] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 609.844294] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e06cd5-8d49-ea01-6612-ffcbc6a61212" [ 609.844294] env[65385]: _type = "Task" [ 609.844294] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.850168] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453104, 'name': ReconfigVM_Task, 'duration_secs': 0.329631} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.851178] env[65385]: DEBUG nova.network.neutron [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Successfully created port: 7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 609.857246] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Reconfigured VM instance instance-00000008 to attach disk [datastore1] a8c76329-5056-4cf0-ac32-85d46429d3db/a8c76329-5056-4cf0-ac32-85d46429d3db.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 609.858714] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3371fd3-c1e2-4465-a3d1-d755efd24973 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.864259] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e06cd5-8d49-ea01-6612-ffcbc6a61212, 'name': SearchDatastore_Task, 'duration_secs': 0.010824} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.865142] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.865432] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 609.865675] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.865816] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.866025] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 609.866370] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7be61a7-d8d5-4bd3-a75b-79706ac874c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.870553] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 609.870553] env[65385]: value = "task-4453111" [ 609.870553] env[65385]: _type = "Task" [ 609.870553] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.877825] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 609.877825] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 609.884971] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcb6d357-57ff-4b2b-8de3-87f34ca8fccf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.888819] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453111, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.894102] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 609.894102] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521e0f8b-9146-cec7-7aa5-832d7a9a04b0" [ 609.894102] env[65385]: _type = "Task" [ 609.894102] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.904163] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521e0f8b-9146-cec7-7aa5-832d7a9a04b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.938538] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Releasing lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.938917] env[65385]: DEBUG nova.compute.manager [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Received event network-vif-plugged-a9fd2b9d-1995-43d9-8550-a27e7828eeb6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 609.939346] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Acquiring lock "ef235d8e-6193-4010-ac50-72cd275e965a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.939708] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Lock "ef235d8e-6193-4010-ac50-72cd275e965a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.939983] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Lock "ef235d8e-6193-4010-ac50-72cd275e965a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 609.940261] env[65385]: DEBUG nova.compute.manager [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] No waiting events found dispatching network-vif-plugged-a9fd2b9d-1995-43d9-8550-a27e7828eeb6 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 609.940529] env[65385]: WARNING nova.compute.manager [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Received unexpected event network-vif-plugged-a9fd2b9d-1995-43d9-8550-a27e7828eeb6 for instance with vm_state building and task_state spawning. [ 609.940792] env[65385]: DEBUG nova.compute.manager [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Received event network-changed-a9fd2b9d-1995-43d9-8550-a27e7828eeb6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 609.941030] env[65385]: DEBUG nova.compute.manager [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Refreshing instance network info cache due to event network-changed-a9fd2b9d-1995-43d9-8550-a27e7828eeb6. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 609.941308] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Acquiring lock "refresh_cache-ef235d8e-6193-4010-ac50-72cd275e965a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.941513] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Acquired lock "refresh_cache-ef235d8e-6193-4010-ac50-72cd275e965a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.941937] env[65385]: DEBUG nova.network.neutron [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Refreshing network info cache for port a9fd2b9d-1995-43d9-8550-a27e7828eeb6 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 609.944636] env[65385]: DEBUG nova.compute.manager [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 609.944960] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.945462] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453107, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.947301] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec483ea0-69b0-4d24-8aa5-3ca79468502d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.964695] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 609.965595] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3feaef2-4f4b-42ff-9ef7-fc983c93041c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.975069] env[65385]: DEBUG oslo_vmware.api [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 609.975069] env[65385]: value = "task-4453112" [ 609.975069] env[65385]: _type = "Task" [ 609.975069] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.992458] env[65385]: DEBUG oslo_vmware.api [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453112, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.082568] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453102, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.104988] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453108, 'name': Rename_Task, 'duration_secs': 0.17536} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.105484] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 610.106387] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3f5a08f-6d12-4dfc-ad0c-905e2965f425 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.114859] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 610.114859] env[65385]: value = "task-4453113" [ 610.114859] env[65385]: _type = "Task" [ 610.114859] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.129321] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453113, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.201439] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453109, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.213690] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4798b68e-922e-49d1-b20c-e989dc419c65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.228660] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453110, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.232474] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a740be-3fad-4eed-bb90-a48e073265a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.271604] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a1af9e-aa44-45ef-b0a1-d0f300d40bee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.276625] env[65385]: DEBUG nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 610.285507] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff359b6-a804-4be9-8eb1-602f0188c4c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.301366] env[65385]: DEBUG nova.compute.provider_tree [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.311872] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 610.312161] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 610.312332] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 610.312516] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 610.312921] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 610.312921] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 610.313067] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.313137] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 610.313298] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 610.313452] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 610.313616] env[65385]: DEBUG nova.virt.hardware [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 610.314575] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5c3bf5-aad4-45ca-a435-f67716aabb05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.324596] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b50dc83-d850-41dd-9b7e-9b6590d3a4dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.381355] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453111, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.405249] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521e0f8b-9146-cec7-7aa5-832d7a9a04b0, 'name': SearchDatastore_Task, 'duration_secs': 0.05882} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.406039] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a031db0e-2250-4cfd-8f89-ea045f259cd6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.412189] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 610.412189] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524592bb-f725-d133-3840-1ee4c9a5852b" [ 610.412189] env[65385]: _type = "Task" [ 610.412189] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.421819] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524592bb-f725-d133-3840-1ee4c9a5852b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.427492] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453107, 'name': CreateVM_Task, 'duration_secs': 0.657587} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.427492] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 610.427838] env[65385]: WARNING neutronclient.v2_0.client [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 610.428580] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.428580] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.428708] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 610.428875] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98f03df6-3a6a-4f40-8c26-a8c9494cb9e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.433998] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 610.433998] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529848ce-1654-11d9-d14a-09d7ffa8dac2" [ 610.433998] env[65385]: _type = "Task" [ 610.433998] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.441947] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529848ce-1654-11d9-d14a-09d7ffa8dac2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.446550] env[65385]: WARNING neutronclient.v2_0.client [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 610.447285] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.447525] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.489379] env[65385]: DEBUG oslo_vmware.api [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453112, 'name': PowerOffVM_Task, 'duration_secs': 0.31402} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.489691] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 610.489783] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 610.490064] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b938137-bf9b-4399-8daa-5d621c55a770 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.555222] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 610.555460] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 610.555716] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Deleting the datastore file [datastore2] b72885c3-5146-42a5-82e8-444cfd89413d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 610.556109] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa899d81-a4d8-4595-96f1-08ddb9ebb976 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.563946] env[65385]: DEBUG oslo_vmware.api [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for the task: (returnval){ [ 610.563946] env[65385]: value = "task-4453115" [ 610.563946] env[65385]: _type = "Task" [ 610.563946] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.573408] env[65385]: DEBUG oslo_vmware.api [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.582109] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453102, 'name': CreateVM_Task, 'duration_secs': 1.140266} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.582277] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 610.582676] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.582822] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.583142] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 610.583387] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4b54d5e-523a-41d9-9ee7-2827f25f7999 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.589018] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 610.589018] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5264196c-ebed-891d-c6df-42e02475b1de" [ 610.589018] env[65385]: _type = "Task" [ 610.589018] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.598648] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5264196c-ebed-891d-c6df-42e02475b1de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.630710] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453113, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.696659] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453109, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.726703] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453110, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544261} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.726969] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] bf79fbcd-e943-4ff3-bcf8-6a2817df8292/bf79fbcd-e943-4ff3-bcf8-6a2817df8292.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 610.727268] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 610.727513] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a82253c7-0103-4d33-b4f8-202deddd2959 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.736303] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 610.736303] env[65385]: value = "task-4453116" [ 610.736303] env[65385]: _type = "Task" [ 610.736303] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.746168] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453116, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.807235] env[65385]: DEBUG nova.scheduler.client.report [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 610.887592] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453111, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.926253] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524592bb-f725-d133-3840-1ee4c9a5852b, 'name': SearchDatastore_Task, 'duration_secs': 0.010512} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.926253] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 610.926253] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] c5b85574-df14-4a7a-ada0-c9668bf18a33/c5b85574-df14-4a7a-ada0-c9668bf18a33.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 610.926253] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a56344c1-7fde-4763-aa32-04daec5a4d23 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.940684] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 610.940684] env[65385]: value = "task-4453117" [ 610.940684] env[65385]: _type = "Task" [ 610.940684] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.956166] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529848ce-1654-11d9-d14a-09d7ffa8dac2, 'name': SearchDatastore_Task, 'duration_secs': 0.01607} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.956166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 610.956166] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 610.956166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.956462] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.956462] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 610.957352] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9976134b-7211-4ff2-a0d0-6597d9c81a9e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.964834] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.964834] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.971695] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.981016] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 610.981255] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 610.982240] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2983875f-595a-4310-8174-1701457161c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.989614] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 610.989614] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5285d013-ec9f-fd6d-0eb1-794721c2a112" [ 610.989614] env[65385]: _type = "Task" [ 610.989614] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.999218] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5285d013-ec9f-fd6d-0eb1-794721c2a112, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.075060] env[65385]: DEBUG oslo_vmware.api [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Task: {'id': task-4453115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265437} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.076311] env[65385]: WARNING neutronclient.v2_0.client [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.076904] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.077274] env[65385]: WARNING openstack [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.085272] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 611.085465] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 611.085629] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.085833] env[65385]: INFO nova.compute.manager [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 611.086170] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 611.086620] env[65385]: DEBUG nova.compute.manager [-] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 611.086718] env[65385]: DEBUG nova.network.neutron [-] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 611.086945] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.087516] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.087778] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.106884] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5264196c-ebed-891d-c6df-42e02475b1de, 'name': SearchDatastore_Task, 'duration_secs': 0.016412} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.106884] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 611.106884] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 611.106884] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.107336] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 611.107336] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 611.107336] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bdf74d92-0aca-4261-945a-4b7ab17edb8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.127351] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 611.127351] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 611.128019] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b830fb31-6273-4790-afa8-f21760612849 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.134564] env[65385]: DEBUG oslo_vmware.api [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453113, 'name': PowerOnVM_Task, 'duration_secs': 0.577099} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.135250] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 611.135480] env[65385]: INFO nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Took 14.60 seconds to spawn the instance on the hypervisor. [ 611.135651] env[65385]: DEBUG nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 611.136463] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d31535a-6451-4d44-a342-025234739ca0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.140723] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 611.140723] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c54ab-f9b8-efa4-efc7-4efa5740278b" [ 611.140723] env[65385]: _type = "Task" [ 611.140723] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.158711] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c54ab-f9b8-efa4-efc7-4efa5740278b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.199802] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453109, 'name': Rename_Task, 'duration_secs': 1.181086} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.201037] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 611.201321] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-366fe541-acce-4217-8dab-3a6761231cc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.209985] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 611.209985] env[65385]: value = "task-4453118" [ 611.209985] env[65385]: _type = "Task" [ 611.209985] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.219158] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453118, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.248169] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453116, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068922} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.248501] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 611.249311] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b98cf9-ce36-4286-86b9-779631f40ee4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.287042] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] bf79fbcd-e943-4ff3-bcf8-6a2817df8292/bf79fbcd-e943-4ff3-bcf8-6a2817df8292.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 611.289328] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9be650ce-4fc4-40f2-909c-7981d98bd074 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.309937] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 611.309937] env[65385]: value = "task-4453119" [ 611.309937] env[65385]: _type = "Task" [ 611.309937] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.320276] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 611.320882] env[65385]: DEBUG nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 611.323554] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.324186] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.503s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 611.326030] env[65385]: INFO nova.compute.claims [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.357486] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.357689] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 611.357903] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.358096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 611.358255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 611.370123] env[65385]: INFO nova.compute.manager [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Terminating instance [ 611.383103] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453111, 'name': Rename_Task, 'duration_secs': 1.095933} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.383363] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 611.383636] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3000fe2-9ef8-4590-8bbc-9f1063b2e607 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.392521] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 611.392521] env[65385]: value = "task-4453120" [ 611.392521] env[65385]: _type = "Task" [ 611.392521] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.402443] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453120, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.416599] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.452485] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453117, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.498768] env[65385]: DEBUG nova.network.neutron [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Updated VIF entry in instance network info cache for port a9fd2b9d-1995-43d9-8550-a27e7828eeb6. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 611.499213] env[65385]: DEBUG nova.network.neutron [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Updating instance_info_cache with network_info: [{"id": "a9fd2b9d-1995-43d9-8550-a27e7828eeb6", "address": "fa:16:3e:6d:ee:ef", "network": {"id": "c9801743-abb3-4791-8958-57966529d5ac", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-267718441-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "66d654f07ad84cd580f841a7fd31bb9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8f580e6-1d86-41ee-9ebe-c531cb9299c6", "external-id": "nsx-vlan-transportzone-150", "segmentation_id": 150, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9fd2b9d-19", "ovs_interfaceid": "a9fd2b9d-1995-43d9-8550-a27e7828eeb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 611.507499] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5285d013-ec9f-fd6d-0eb1-794721c2a112, 'name': SearchDatastore_Task, 'duration_secs': 0.056369} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.510648] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cb05645-7e2f-4afc-b824-cd68614f9583 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.518414] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 611.518414] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c4abf1-863f-1846-2db6-61395ba25f54" [ 611.518414] env[65385]: _type = "Task" [ 611.518414] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.529537] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c4abf1-863f-1846-2db6-61395ba25f54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.635359] env[65385]: DEBUG nova.network.neutron [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Successfully updated port: 7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 611.660387] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c54ab-f9b8-efa4-efc7-4efa5740278b, 'name': SearchDatastore_Task, 'duration_secs': 0.02025} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.667613] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-540cdfe7-af08-405f-84d7-c621a500c45f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.671516] env[65385]: INFO nova.compute.manager [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Took 29.42 seconds to build instance. [ 611.679236] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 611.679236] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ad5045-1f01-d8ae-0206-6947f91daca7" [ 611.679236] env[65385]: _type = "Task" [ 611.679236] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.691967] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ad5045-1f01-d8ae-0206-6947f91daca7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.723635] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453118, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.796477] env[65385]: DEBUG nova.compute.manager [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Received event network-changed-6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 611.796477] env[65385]: DEBUG nova.compute.manager [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Refreshing instance network info cache due to event network-changed-6b5cfde5-8f86-4000-90b9-5c6140ea1606. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 611.796477] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Acquiring lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.796477] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Acquired lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 611.796477] env[65385]: DEBUG nova.network.neutron [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Refreshing network info cache for port 6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 611.829974] env[65385]: DEBUG nova.compute.utils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 611.831654] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.832486] env[65385]: DEBUG nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 611.833230] env[65385]: DEBUG nova.network.neutron [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 611.833230] env[65385]: WARNING neutronclient.v2_0.client [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.833364] env[65385]: WARNING neutronclient.v2_0.client [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.833952] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.836291] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.874071] env[65385]: DEBUG nova.compute.manager [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 611.874071] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.874071] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f94ee9-7f6d-45dd-a2c3-8856064f880b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.886996] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 611.886996] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67711573-9e8b-49ff-9c85-2dfea628b0ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.898371] env[65385]: DEBUG oslo_vmware.api [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 611.898371] env[65385]: value = "task-4453121" [ 611.898371] env[65385]: _type = "Task" [ 611.898371] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.907445] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453120, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.911748] env[65385]: DEBUG nova.policy [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d5103a5769a4296973351194f49fed3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '754b1facaaa14501b2204c98e1d7a5a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 611.920967] env[65385]: DEBUG oslo_vmware.api [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.956981] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453117, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.900563} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.956981] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] c5b85574-df14-4a7a-ada0-c9668bf18a33/c5b85574-df14-4a7a-ada0-c9668bf18a33.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 611.957579] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 611.957579] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3babd2d6-b9a9-4ec0-89bf-3bd16b453b6b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.968036] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 611.968036] env[65385]: value = "task-4453122" [ 611.968036] env[65385]: _type = "Task" [ 611.968036] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.979487] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453122, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.009040] env[65385]: DEBUG oslo_concurrency.lockutils [req-f129d875-541f-4865-8343-078b2acbd81a req-a01dc931-df09-458c-81cd-85ac1e40d85d service nova] Releasing lock "refresh_cache-ef235d8e-6193-4010-ac50-72cd275e965a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.034045] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c4abf1-863f-1846-2db6-61395ba25f54, 'name': SearchDatastore_Task, 'duration_secs': 0.057025} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.034703] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.036144] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8ac44124-d8cf-421c-9ae4-c943df94550a/8ac44124-d8cf-421c-9ae4-c943df94550a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 612.036144] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e209495-b503-45f5-9d2d-c117a5bc17cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.048023] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 612.048023] env[65385]: value = "task-4453123" [ 612.048023] env[65385]: _type = "Task" [ 612.048023] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.059593] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.141699] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.141699] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 612.141699] env[65385]: DEBUG nova.network.neutron [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 612.175044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae45e7d0-b5b0-4d33-a52f-7425fb9dddf3 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.935s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.202029] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ad5045-1f01-d8ae-0206-6947f91daca7, 'name': SearchDatastore_Task, 'duration_secs': 0.037711} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.202029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.202029] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 612.202029] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52524e59-df39-41b9-a9f7-73a158450f45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.210142] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 612.210142] env[65385]: value = "task-4453124" [ 612.210142] env[65385]: _type = "Task" [ 612.210142] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.225183] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.232052] env[65385]: DEBUG oslo_vmware.api [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453118, 'name': PowerOnVM_Task, 'duration_secs': 0.661655} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.232258] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 612.232445] env[65385]: INFO nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Took 13.25 seconds to spawn the instance on the hypervisor. [ 612.232670] env[65385]: DEBUG nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 612.233634] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93834c9-0fcf-45d4-b7e2-2e98746399d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.264091] env[65385]: DEBUG nova.compute.manager [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Received event network-vif-plugged-7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 612.265443] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Acquiring lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.265443] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.265443] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.265443] env[65385]: DEBUG nova.compute.manager [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] No waiting events found dispatching network-vif-plugged-7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 612.265443] env[65385]: WARNING nova.compute.manager [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Received unexpected event network-vif-plugged-7d3f8bf5-22d2-41b0-9743-0616daaef615 for instance with vm_state building and task_state spawning. [ 612.265747] env[65385]: DEBUG nova.compute.manager [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Received event network-changed-7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 612.265747] env[65385]: DEBUG nova.compute.manager [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Refreshing instance network info cache due to event network-changed-7d3f8bf5-22d2-41b0-9743-0616daaef615. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 612.265747] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Acquiring lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.265747] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Acquired lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 612.266147] env[65385]: DEBUG nova.network.neutron [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Refreshing network info cache for port 7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 612.304212] env[65385]: WARNING neutronclient.v2_0.client [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.304298] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.304634] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.319352] env[65385]: DEBUG nova.network.neutron [-] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 612.324431] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453119, 'name': ReconfigVM_Task, 'duration_secs': 0.901754} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.325034] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Reconfigured VM instance instance-00000009 to attach disk [datastore1] bf79fbcd-e943-4ff3-bcf8-6a2817df8292/bf79fbcd-e943-4ff3-bcf8-6a2817df8292.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 612.325707] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9bd700e-7604-4c6c-8ab3-bc2dad376392 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.332889] env[65385]: DEBUG nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 612.338804] env[65385]: DEBUG nova.network.neutron [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Successfully created port: 34c764b8-f45c-4237-9998-f49b871bb7d4 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 612.342029] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 612.342029] env[65385]: value = "task-4453125" [ 612.342029] env[65385]: _type = "Task" [ 612.342029] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.356871] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453125, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.409223] env[65385]: DEBUG oslo_vmware.api [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453120, 'name': PowerOnVM_Task, 'duration_secs': 0.597598} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.409949] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 612.410173] env[65385]: INFO nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Took 11.02 seconds to spawn the instance on the hypervisor. [ 612.410368] env[65385]: DEBUG nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 612.411298] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4ed1c2-0ad1-4117-b409-e395deb4d44d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.421129] env[65385]: DEBUG oslo_vmware.api [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453121, 'name': PowerOffVM_Task, 'duration_secs': 0.353254} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.428044] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 612.428044] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 612.428044] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-285b8806-1921-426b-b7a9-169ab4e1ce00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.477295] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.477714] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.489946] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453122, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07863} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.492926] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 612.494845] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fcefa29-9378-4f16-a226-0f16e276bdbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.522330] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] c5b85574-df14-4a7a-ada0-c9668bf18a33/c5b85574-df14-4a7a-ada0-c9668bf18a33.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 612.532373] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c54143db-a0d0-434f-83d4-7f2630e73a5a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.553377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 612.553377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 612.553377] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Deleting the datastore file [datastore2] 6e1eeff2-1587-44c6-a154-f73a2d0f8061 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 612.554832] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78908d0e-7b87-4084-8d23-09ae0881d542 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.567864] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453123, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.573393] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 612.573393] env[65385]: value = "task-4453128" [ 612.573393] env[65385]: _type = "Task" [ 612.573393] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.573695] env[65385]: DEBUG oslo_vmware.api [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for the task: (returnval){ [ 612.573695] env[65385]: value = "task-4453127" [ 612.573695] env[65385]: _type = "Task" [ 612.573695] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.588238] env[65385]: DEBUG oslo_vmware.api [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453127, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.594840] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453128, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.596565] env[65385]: WARNING neutronclient.v2_0.client [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.597267] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.597267] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.646367] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.646367] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.685618] env[65385]: DEBUG nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 612.722674] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453124, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.759671] env[65385]: DEBUG nova.network.neutron [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 612.765012] env[65385]: INFO nova.compute.manager [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Took 27.73 seconds to build instance. [ 612.770031] env[65385]: WARNING neutronclient.v2_0.client [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.770031] env[65385]: WARNING openstack [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.770238] env[65385]: WARNING openstack [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.811345] env[65385]: DEBUG nova.network.neutron [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Updated VIF entry in instance network info cache for port 6b5cfde5-8f86-4000-90b9-5c6140ea1606. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 612.811848] env[65385]: DEBUG nova.network.neutron [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Updating instance_info_cache with network_info: [{"id": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "address": "fa:16:3e:77:44:52", "network": {"id": "8919702d-2cd0-4d13-8f2f-31653142115a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-363641611-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a894ba2e7f7c42d091184408d0b84343", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5cfde5-8f", "ovs_interfaceid": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 612.816401] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81bbf4b-87aa-4b37-9896-8eb251e74e9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.825914] env[65385]: INFO nova.compute.manager [-] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Took 1.74 seconds to deallocate network for instance. [ 612.835035] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c655d2-0d01-4b93-9092-a192cd49bf03 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.878716] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9313edd1-67e6-4081-a65c-b85d05160bea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.890520] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453125, 'name': Rename_Task, 'duration_secs': 0.210061} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.893815] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 612.894164] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65c192ee-eb2c-4e0c-b5f0-daea514307b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.897057] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bd80dc-02b7-40ae-b80b-7b3286b59619 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.918784] env[65385]: DEBUG nova.compute.provider_tree [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.921731] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 612.921731] env[65385]: value = "task-4453129" [ 612.921731] env[65385]: _type = "Task" [ 612.921731] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.935060] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.936953] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.937538] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.954339] env[65385]: INFO nova.compute.manager [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Took 24.64 seconds to build instance. [ 613.068818] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453123, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.74981} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.073457] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8ac44124-d8cf-421c-9ae4-c943df94550a/8ac44124-d8cf-421c-9ae4-c943df94550a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 613.073731] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 613.074296] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54226d7f-a0be-411d-90a8-9808b6f32916 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.089656] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.089899] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.094027] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 613.094027] env[65385]: value = "task-4453130" [ 613.094027] env[65385]: _type = "Task" [ 613.094027] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.102127] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453128, 'name': ReconfigVM_Task, 'duration_secs': 0.370693} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.102396] env[65385]: DEBUG oslo_vmware.api [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453127, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.106500] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Reconfigured VM instance instance-0000000b to attach disk [datastore1] c5b85574-df14-4a7a-ada0-c9668bf18a33/c5b85574-df14-4a7a-ada0-c9668bf18a33.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 613.107827] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98ba8485-4c08-4e20-b53c-ab4a35c832c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.116892] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.123292] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 613.123292] env[65385]: value = "task-4453131" [ 613.123292] env[65385]: _type = "Task" [ 613.123292] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.134650] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453131, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.169211] env[65385]: WARNING neutronclient.v2_0.client [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.170071] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.170503] env[65385]: WARNING openstack [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.212779] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.227516] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453124, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.941241} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.229568] env[65385]: WARNING openstack [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.230188] env[65385]: WARNING openstack [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.240546] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 613.240842] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 613.241765] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e655ef7-b641-4b8f-be17-df89ed62608c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.249455] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 613.249455] env[65385]: value = "task-4453132" [ 613.249455] env[65385]: _type = "Task" [ 613.249455] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.261484] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453132, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.280278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1949a190-536d-4161-a21a-9656658881c3 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "ef235d8e-6193-4010-ac50-72cd275e965a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.261s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.317403] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Releasing lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 613.318570] env[65385]: DEBUG nova.compute.manager [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Received event network-vif-plugged-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 613.318864] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Acquiring lock "8ac44124-d8cf-421c-9ae4-c943df94550a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.319327] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.319327] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.320694] env[65385]: DEBUG nova.compute.manager [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] No waiting events found dispatching network-vif-plugged-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 613.320694] env[65385]: WARNING nova.compute.manager [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Received unexpected event network-vif-plugged-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d for instance with vm_state building and task_state spawning. [ 613.320694] env[65385]: DEBUG nova.compute.manager [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Received event network-changed-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 613.320694] env[65385]: DEBUG nova.compute.manager [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Refreshing instance network info cache due to event network-changed-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 613.320694] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Acquiring lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.320969] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Acquired lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 613.320969] env[65385]: DEBUG nova.network.neutron [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Refreshing network info cache for port f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 613.333429] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.376415] env[65385]: DEBUG nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 613.407404] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 613.407690] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 613.407858] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 613.408165] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 613.408281] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 613.408411] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 613.408706] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.408916] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 613.409143] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 613.409316] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 613.409713] env[65385]: DEBUG nova.virt.hardware [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 613.410515] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865da8d2-a7cb-47ef-8513-7efadd0ebda1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.421833] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0ccf33-164b-4dec-9c57-efa0ebed8ed8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.427520] env[65385]: DEBUG nova.scheduler.client.report [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 613.445425] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453129, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.459401] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf1bcd64-1f14-4e74-953e-4a5969a02b40 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.151s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.588949] env[65385]: DEBUG oslo_vmware.api [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Task: {'id': task-4453127, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.670859} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.588949] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 613.588949] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 613.588949] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.589207] env[65385]: INFO nova.compute.manager [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Took 1.72 seconds to destroy the instance on the hypervisor. [ 613.589665] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 613.589665] env[65385]: DEBUG nova.compute.manager [-] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 613.589665] env[65385]: DEBUG nova.network.neutron [-] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 613.589877] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.590415] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.590681] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.612605] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082584} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.612755] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 613.613860] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c46f525-fbdb-4faa-a038-73073a78e957 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.638345] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] 8ac44124-d8cf-421c-9ae4-c943df94550a/8ac44124-d8cf-421c-9ae4-c943df94550a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 613.643023] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d91e642-cd3d-4eae-9ecc-88c429766c2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.668851] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453131, 'name': Rename_Task, 'duration_secs': 0.203225} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.673250] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 613.673250] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 613.673250] env[65385]: value = "task-4453133" [ 613.673250] env[65385]: _type = "Task" [ 613.673250] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.673250] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-737ad636-4677-4679-99c3-347132cc2dbf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.681863] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453133, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.683711] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 613.683711] env[65385]: value = "task-4453134" [ 613.683711] env[65385]: _type = "Task" [ 613.683711] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.692872] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453134, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.756838] env[65385]: DEBUG nova.network.neutron [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Updating instance_info_cache with network_info: [{"id": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "address": "fa:16:3e:5c:e2:96", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a4f4b0b-f7", "ovs_interfaceid": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 613.764968] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453132, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12721} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.765330] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 613.766506] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc3c3f4-81b0-4806-9ebd-1f30fc452ed1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.783692] env[65385]: DEBUG nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 613.796972] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 613.798726] env[65385]: WARNING neutronclient.v2_0.client [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.799810] env[65385]: WARNING openstack [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.800264] env[65385]: WARNING openstack [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.809551] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50a8642b-e520-4961-afba-529e3a44165e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.825915] env[65385]: WARNING neutronclient.v2_0.client [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.826579] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.827187] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.840771] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 613.840771] env[65385]: value = "task-4453135" [ 613.840771] env[65385]: _type = "Task" [ 613.840771] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.849590] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.858548] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453135, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.933881] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.934028] env[65385]: DEBUG nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 613.939805] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.398s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.941227] env[65385]: INFO nova.compute.claims [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.944026] env[65385]: DEBUG oslo_vmware.api [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453129, 'name': PowerOnVM_Task, 'duration_secs': 0.876458} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.944604] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 613.944815] env[65385]: INFO nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Took 10.64 seconds to spawn the instance on the hypervisor. [ 613.945012] env[65385]: DEBUG nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 613.945834] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7665f43-3f69-40d0-ad2c-fba8e660b7e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.965553] env[65385]: DEBUG nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 614.048994] env[65385]: DEBUG nova.network.neutron [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Successfully updated port: 34c764b8-f45c-4237-9998-f49b871bb7d4 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 614.184724] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.197695] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453134, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.260611] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 614.261013] env[65385]: DEBUG nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Instance network_info: |[{"id": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "address": "fa:16:3e:5c:e2:96", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a4f4b0b-f7", "ovs_interfaceid": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 614.261523] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:e2:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7514a465-f1a4-4a8b-b76b-726b1a9d7e2f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 614.273986] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Creating folder: Project (112402342b0e481f83be7e17c9f739e9). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 614.274403] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d344258-0b3d-499c-b2fc-c76ee0114bd7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.289533] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Created folder: Project (112402342b0e481f83be7e17c9f739e9) in parent group-v870881. [ 614.290112] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Creating folder: Instances. Parent ref: group-v870916. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 614.290446] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efb6847f-01d8-4ea3-9ca7-5fffb1be10a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.302286] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Created folder: Instances in parent group-v870916. [ 614.302523] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 614.305376] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 614.305498] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-775d03c8-e21e-48a1-aee7-5aad5a3a17b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.323532] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.328392] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 614.328392] env[65385]: value = "task-4453138" [ 614.328392] env[65385]: _type = "Task" [ 614.328392] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.335810] env[65385]: DEBUG nova.network.neutron [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updated VIF entry in instance network info cache for port 7d3f8bf5-22d2-41b0-9743-0616daaef615. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 614.336166] env[65385]: DEBUG nova.network.neutron [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updating instance_info_cache with network_info: [{"id": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "address": "fa:16:3e:f9:84:04", "network": {"id": "9ed70dba-8022-422b-8af1-9e125786bbe0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-823001245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73c43e9322314241811db1a37e18f9b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f8bf5-22", "ovs_interfaceid": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 614.343143] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453138, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.356207] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.431473] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.431989] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.448316] env[65385]: DEBUG nova.compute.utils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 614.450028] env[65385]: DEBUG nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 614.450249] env[65385]: DEBUG nova.network.neutron [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 614.450627] env[65385]: WARNING neutronclient.v2_0.client [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.450915] env[65385]: WARNING neutronclient.v2_0.client [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.451470] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.452295] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.474390] env[65385]: INFO nova.compute.manager [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Took 23.49 seconds to build instance. [ 614.515136] env[65385]: DEBUG nova.policy [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7e18058b77c4ccaa8afdf15e85930a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e94f2a868e2d4d31af54f2e5d8c479f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 614.518836] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.552061] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-f77d861a-ae66-44c3-901d-e499926d6f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.552165] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-f77d861a-ae66-44c3-901d-e499926d6f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 614.552371] env[65385]: DEBUG nova.network.neutron [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 614.556035] env[65385]: WARNING neutronclient.v2_0.client [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.556200] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.556535] env[65385]: WARNING openstack [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.687360] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453133, 'name': ReconfigVM_Task, 'duration_secs': 0.735714} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.691502] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Reconfigured VM instance instance-0000000a to attach disk [datastore1] 8ac44124-d8cf-421c-9ae4-c943df94550a/8ac44124-d8cf-421c-9ae4-c943df94550a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 614.691706] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07c881e8-8b50-4f7e-abf0-ec734633d386 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.702254] env[65385]: DEBUG oslo_vmware.api [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453134, 'name': PowerOnVM_Task, 'duration_secs': 0.951089} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.703910] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 614.706032] env[65385]: INFO nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Took 6.55 seconds to spawn the instance on the hypervisor. [ 614.706032] env[65385]: DEBUG nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 614.706032] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 614.706032] env[65385]: value = "task-4453139" [ 614.706032] env[65385]: _type = "Task" [ 614.706032] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.706032] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0f1706-f46f-4888-8fe2-9d706ce725a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.730032] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453139, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.846916] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] Releasing lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 614.846916] env[65385]: DEBUG nova.compute.manager [req-2a9bdf03-b800-49d8-a7ed-c804b3079932 req-4142a49b-8119-4ae0-9ce8-e4a4b51f3e11 service nova] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Received event network-vif-deleted-80afc4c9-4d95-4675-a30d-521b87803aa6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 614.846916] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453138, 'name': CreateVM_Task, 'duration_secs': 0.502681} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.850151] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 614.850712] env[65385]: DEBUG nova.network.neutron [-] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 614.852210] env[65385]: WARNING neutronclient.v2_0.client [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.852683] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.852683] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 614.853080] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 614.854025] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0826bfbe-6a3d-4974-becb-5264b4130bd5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.861422] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453135, 'name': ReconfigVM_Task, 'duration_secs': 0.795408} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.862708] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Reconfigured VM instance instance-00000005 to attach disk [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 614.864238] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 614.864238] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5271009a-763c-72bd-53c7-51182b086c53" [ 614.864238] env[65385]: _type = "Task" [ 614.864238] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.864238] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-839c8dff-0512-489c-98cf-7a90ddbe531a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.881307] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5271009a-763c-72bd-53c7-51182b086c53, 'name': SearchDatastore_Task, 'duration_secs': 0.01693} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.883907] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 614.884480] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 614.884738] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.884887] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 614.885078] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 614.885943] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 614.885943] env[65385]: value = "task-4453140" [ 614.885943] env[65385]: _type = "Task" [ 614.885943] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.886199] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63e3f4dc-c95f-458b-b0d6-3efc13fcc7ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.901957] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453140, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.907315] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 614.907561] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 614.908402] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5dffb78-cbc9-4030-8d37-468c34fddd14 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.915090] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 614.915090] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1e2f8-3ebc-904b-79d7-e7675bf42afc" [ 614.915090] env[65385]: _type = "Task" [ 614.915090] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.928840] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1e2f8-3ebc-904b-79d7-e7675bf42afc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.963309] env[65385]: DEBUG nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 614.978476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b0c15e6-8ca1-4bc9-9e5e-d6d3dc1840c4 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.005s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.029836] env[65385]: DEBUG nova.network.neutron [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Successfully created port: e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 615.055625] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.055988] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.083595] env[65385]: DEBUG nova.network.neutron [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Updated VIF entry in instance network info cache for port f5dabab4-c5af-42f4-bfdc-eec4330bdb0d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 615.084554] env[65385]: DEBUG nova.network.neutron [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Updating instance_info_cache with network_info: [{"id": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "address": "fa:16:3e:0e:a3:79", "network": {"id": "aec9c0ce-a40e-4d0b-ad1d-2984fe3a4e89", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1834228009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "349d3e2d0d8849819007c19528145c3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5dabab4-c5", "ovs_interfaceid": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 615.227372] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453139, 'name': Rename_Task, 'duration_secs': 0.195028} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.232301] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 615.236206] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00a59011-cda5-449e-84b6-8809b7619264 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.237945] env[65385]: INFO nova.compute.manager [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Took 17.02 seconds to build instance. [ 615.248469] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 615.248469] env[65385]: value = "task-4453141" [ 615.248469] env[65385]: _type = "Task" [ 615.248469] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.261722] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453141, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.331808] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fafd88-baf2-43dd-86de-dee72df9fc3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.341037] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89179ce-8bf5-45c1-877d-c1b3bfd25fd3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.375386] env[65385]: INFO nova.compute.manager [-] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Took 1.79 seconds to deallocate network for instance. [ 615.378630] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6dc4f0b-fc69-4ad2-97a1-1bb8521a8639 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.394946] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6988eaa-076f-4c88-8817-d62076f06bbe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.410639] env[65385]: DEBUG nova.compute.provider_tree [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.418245] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453140, 'name': Rename_Task, 'duration_secs': 0.298729} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.422126] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 615.423112] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-750a48ad-d410-40e1-92b9-1a0dd299a1e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.433307] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1e2f8-3ebc-904b-79d7-e7675bf42afc, 'name': SearchDatastore_Task, 'duration_secs': 0.024986} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.436187] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 615.436187] env[65385]: value = "task-4453142" [ 615.436187] env[65385]: _type = "Task" [ 615.436187] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.436513] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ba0bead-0ff2-4a18-9139-d48734576ba8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.451512] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 615.451512] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529b1a3d-aae7-5d62-5ce1-6543b339a785" [ 615.451512] env[65385]: _type = "Task" [ 615.451512] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.451829] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453142, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.461509] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529b1a3d-aae7-5d62-5ce1-6543b339a785, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.484428] env[65385]: DEBUG nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 615.553384] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "eb9d0510-b453-4695-9e1d-731217b9f8ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 615.553617] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.587492] env[65385]: DEBUG oslo_concurrency.lockutils [req-1699c40a-8d7c-4d65-9549-a9c9f4de6d93 req-fe232428-a04d-4ffc-88af-70df73f8d292 service nova] Releasing lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 615.664321] env[65385]: DEBUG nova.network.neutron [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 615.741055] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ab4aef00-9ce8-4f01-a4de-20ab916c30c6 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "c5b85574-df14-4a7a-ada0-c9668bf18a33" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.414s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.762539] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453141, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.885692] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 615.922777] env[65385]: DEBUG nova.scheduler.client.report [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 615.951199] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453142, 'name': PowerOnVM_Task} progress is 81%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.961928] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529b1a3d-aae7-5d62-5ce1-6543b339a785, 'name': SearchDatastore_Task, 'duration_secs': 0.024532} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.962242] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 615.962440] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 649b4e21-fc55-415c-a1f3-ec724397b874/649b4e21-fc55-415c-a1f3-ec724397b874.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 615.962896] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b70a5b5-bf69-44ef-bcb5-f059bf8d9039 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.970776] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 615.970776] env[65385]: value = "task-4453143" [ 615.970776] env[65385]: _type = "Task" [ 615.970776] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.976238] env[65385]: DEBUG nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 615.985300] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.004225] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 616.005336] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 616.005525] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 616.005735] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 616.005878] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 616.006046] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 616.006299] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.006464] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 616.006659] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 616.006833] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 616.007034] env[65385]: DEBUG nova.virt.hardware [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 616.008298] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91ce7ca-04a1-4af1-9f81-e7b62efd99cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.011716] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.018796] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20232904-a59e-48ae-a6cf-0247d68361c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.178072] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.178764] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.246037] env[65385]: DEBUG nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 616.261939] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453141, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.427121] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 616.427746] env[65385]: DEBUG nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 616.430818] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.117s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.431278] env[65385]: DEBUG nova.objects.instance [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Lazy-loading 'resources' on Instance uuid 504f5660-6715-4c7a-965b-9d2ef1852391 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 616.455828] env[65385]: DEBUG oslo_vmware.api [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453142, 'name': PowerOnVM_Task, 'duration_secs': 0.84705} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.456095] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.456486] env[65385]: DEBUG nova.compute.manager [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 616.457287] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2f1d4a-467f-4bbc-8fbe-8bd681635f6b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.484555] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453143, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.766291] env[65385]: DEBUG oslo_vmware.api [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453141, 'name': PowerOnVM_Task, 'duration_secs': 1.071309} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.766553] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.766788] env[65385]: INFO nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Took 11.08 seconds to spawn the instance on the hypervisor. [ 616.766995] env[65385]: DEBUG nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 616.767916] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9516c9b1-1220-42dc-be2d-d0e09121fed8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.782157] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.840853] env[65385]: DEBUG nova.network.neutron [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Successfully updated port: e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 616.858324] env[65385]: WARNING neutronclient.v2_0.client [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.859361] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.859361] env[65385]: WARNING openstack [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.935787] env[65385]: DEBUG nova.compute.utils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 616.942189] env[65385]: DEBUG nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 616.942662] env[65385]: DEBUG nova.network.neutron [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 616.944014] env[65385]: WARNING neutronclient.v2_0.client [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.944014] env[65385]: WARNING neutronclient.v2_0.client [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.944014] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.944654] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.952500] env[65385]: DEBUG nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 616.983998] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593717} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.985412] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 649b4e21-fc55-415c-a1f3-ec724397b874/649b4e21-fc55-415c-a1f3-ec724397b874.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 616.985615] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 616.988267] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da51bbf7-d90b-40d5-8d5c-aa2fce9da6fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.990827] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.998055] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 616.998055] env[65385]: value = "task-4453144" [ 616.998055] env[65385]: _type = "Task" [ 616.998055] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.013122] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.296245] env[65385]: INFO nova.compute.manager [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Took 24.98 seconds to build instance. [ 617.344587] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.346577] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.346577] env[65385]: DEBUG nova.network.neutron [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 617.351133] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0948ab-e409-410d-b3ae-90dc3c551c32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.362400] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049cb1b1-c10c-480f-a183-a1acb1c0488f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.402815] env[65385]: DEBUG nova.policy [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53765f0467d14fc3bc946cc893f75280', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '112402342b0e481f83be7e17c9f739e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 617.405343] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57e01aa-b30d-4af2-b039-953ff6f61768 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.415025] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d7a0c4-2a05-427b-8ef9-bf9cf20d396d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.433395] env[65385]: DEBUG nova.compute.provider_tree [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.437136] env[65385]: DEBUG nova.network.neutron [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Updating instance_info_cache with network_info: [{"id": "34c764b8-f45c-4237-9998-f49b871bb7d4", "address": "fa:16:3e:ca:cd:bf", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34c764b8-f4", "ovs_interfaceid": "34c764b8-f45c-4237-9998-f49b871bb7d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 617.511825] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177403} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.512245] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 617.514653] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44a6f4e-214b-4d78-8e4b-9ef8837e5640 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.540751] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Reconfiguring VM instance instance-0000000c to attach disk [datastore2] 649b4e21-fc55-415c-a1f3-ec724397b874/649b4e21-fc55-415c-a1f3-ec724397b874.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 617.541555] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8097962c-c09c-4f70-8579-c10e70919616 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.566146] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 617.566146] env[65385]: value = "task-4453145" [ 617.566146] env[65385]: _type = "Task" [ 617.566146] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.576796] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453145, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.636394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.636394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.636394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.636394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.636716] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.638738] env[65385]: INFO nova.compute.manager [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Terminating instance [ 617.800080] env[65385]: DEBUG oslo_concurrency.lockutils [None req-33a7ebf7-34d8-4e0e-98bc-90740f72be65 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.499s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.838252] env[65385]: DEBUG nova.network.neutron [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Successfully created port: 99c91233-f17e-45f0-8852-fd19e18a8502 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 617.856397] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.856787] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.939648] env[65385]: DEBUG nova.scheduler.client.report [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 617.943083] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-f77d861a-ae66-44c3-901d-e499926d6f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.943407] env[65385]: DEBUG nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Instance network_info: |[{"id": "34c764b8-f45c-4237-9998-f49b871bb7d4", "address": "fa:16:3e:ca:cd:bf", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34c764b8-f4", "ovs_interfaceid": "34c764b8-f45c-4237-9998-f49b871bb7d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 617.944126] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:cd:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34c764b8-f45c-4237-9998-f49b871bb7d4', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 617.952271] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating folder: Project (754b1facaaa14501b2204c98e1d7a5a8). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.952578] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a433b7f-0b4e-4a9b-bcb8-8e365719d9c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.966076] env[65385]: DEBUG nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 617.969378] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created folder: Project (754b1facaaa14501b2204c98e1d7a5a8) in parent group-v870881. [ 617.969584] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating folder: Instances. Parent ref: group-v870919. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.969802] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-845cd4a3-6d84-4336-89df-96fe93139a13 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.981855] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created folder: Instances in parent group-v870919. [ 617.982285] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 617.982437] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 617.982679] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63f6e6a0-2dbf-42d4-9bf9-d0180b40293c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.008296] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 618.008668] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 618.008866] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 618.009114] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 618.009299] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 618.009484] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 618.009762] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.010031] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 618.010190] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 618.010425] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 618.010668] env[65385]: DEBUG nova.virt.hardware [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 618.012036] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb54a07b-323a-45f8-897a-4a9bbd7b08f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.017703] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 618.017703] env[65385]: value = "task-4453148" [ 618.017703] env[65385]: _type = "Task" [ 618.017703] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.028204] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7092ba04-3995-415a-acca-05a940a97977 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.038760] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453148, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.077299] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453145, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.135537] env[65385]: DEBUG nova.network.neutron [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 618.144746] env[65385]: DEBUG nova.compute.manager [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 618.144987] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.145873] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8707bab-072d-4c59-b053-26c8eba4989b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.156221] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 618.156447] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e85ee26-b5b8-4ef7-8a64-349fc3d924a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.169021] env[65385]: DEBUG oslo_vmware.api [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 618.169021] env[65385]: value = "task-4453149" [ 618.169021] env[65385]: _type = "Task" [ 618.169021] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.182730] env[65385]: DEBUG oslo_vmware.api [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.245443] env[65385]: DEBUG nova.compute.manager [req-e4083490-6f54-465d-995a-124136ea36cf req-2728d898-8d91-4f10-bfaa-f2ce31f99197 service nova] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Received event network-vif-deleted-2b700413-9737-481b-95bf-92741b9085bd {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 618.445866] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.015s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 618.449355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.236s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.450355] env[65385]: INFO nova.compute.claims [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.478140] env[65385]: INFO nova.scheduler.client.report [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Deleted allocations for instance 504f5660-6715-4c7a-965b-9d2ef1852391 [ 618.529977] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453148, 'name': CreateVM_Task, 'duration_secs': 0.521088} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.530177] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 618.530737] env[65385]: WARNING neutronclient.v2_0.client [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.532081] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.532081] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.532081] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 618.532081] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c235815a-2389-48d9-b3d0-3042b54ff0ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.535584] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.536112] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.546572] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 618.546572] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d65717-406b-21f2-acef-6349b64ae3fc" [ 618.546572] env[65385]: _type = "Task" [ 618.546572] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.557508] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d65717-406b-21f2-acef-6349b64ae3fc, 'name': SearchDatastore_Task, 'duration_secs': 0.01625} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.557790] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.557959] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 618.558175] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.558316] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.558487] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.558752] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e4dc164-b38e-415e-818b-6f8f0918a692 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.572771] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.572969] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 618.574172] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-594dc290-ebc0-4c51-9b7d-d242a255933d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.581286] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453145, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.585695] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 618.585695] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523ced45-06d7-c5eb-a23f-66538ff5a351" [ 618.585695] env[65385]: _type = "Task" [ 618.585695] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.600121] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523ced45-06d7-c5eb-a23f-66538ff5a351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.679304] env[65385]: DEBUG oslo_vmware.api [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453149, 'name': PowerOffVM_Task, 'duration_secs': 0.33276} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.679646] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 618.679828] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 618.680140] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a78b8a21-4b66-4a18-9387-461fa374e17a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.762630] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 618.764469] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 618.764469] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Deleting the datastore file [datastore2] a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 618.768026] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-421195ef-1d8f-45d2-a569-e61b157ad01c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.777322] env[65385]: DEBUG oslo_vmware.api [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for the task: (returnval){ [ 618.777322] env[65385]: value = "task-4453151" [ 618.777322] env[65385]: _type = "Task" [ 618.777322] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.789420] env[65385]: DEBUG oslo_vmware.api [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.804999] env[65385]: DEBUG nova.compute.manager [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Received event network-vif-plugged-7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 618.805518] env[65385]: DEBUG oslo_concurrency.lockutils [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Acquiring lock "649b4e21-fc55-415c-a1f3-ec724397b874-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.805923] env[65385]: DEBUG oslo_concurrency.lockutils [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Lock "649b4e21-fc55-415c-a1f3-ec724397b874-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.809067] env[65385]: DEBUG oslo_concurrency.lockutils [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Lock "649b4e21-fc55-415c-a1f3-ec724397b874-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 618.809067] env[65385]: DEBUG nova.compute.manager [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] No waiting events found dispatching network-vif-plugged-7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 618.809067] env[65385]: WARNING nova.compute.manager [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Received unexpected event network-vif-plugged-7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 for instance with vm_state building and task_state spawning. [ 618.809067] env[65385]: DEBUG nova.compute.manager [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Received event network-changed-7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 618.809067] env[65385]: DEBUG nova.compute.manager [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Refreshing instance network info cache due to event network-changed-7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 618.809534] env[65385]: DEBUG oslo_concurrency.lockutils [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Acquiring lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.809534] env[65385]: DEBUG oslo_concurrency.lockutils [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Acquired lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.809534] env[65385]: DEBUG nova.network.neutron [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Refreshing network info cache for port 7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 618.869293] env[65385]: WARNING neutronclient.v2_0.client [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.870368] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.870835] env[65385]: WARNING openstack [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.994507] env[65385]: DEBUG oslo_concurrency.lockutils [None req-50b875cb-22aa-451b-8366-f5c3ac5a5ede tempest-DeleteServersAdminTestJSON-1022520972 tempest-DeleteServersAdminTestJSON-1022520972-project-admin] Lock "504f5660-6715-4c7a-965b-9d2ef1852391" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.610s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.048375] env[65385]: DEBUG nova.network.neutron [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updating instance_info_cache with network_info: [{"id": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "address": "fa:16:3e:a7:7e:98", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape37bf46a-30", "ovs_interfaceid": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 619.079230] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453145, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.097991] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523ced45-06d7-c5eb-a23f-66538ff5a351, 'name': SearchDatastore_Task, 'duration_secs': 0.01466} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.098901] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-becfc240-6ab9-4d6e-86d2-edbe213e37ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.106417] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 619.106417] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52366b33-1bf8-9dba-e73a-db36f3f4d10b" [ 619.106417] env[65385]: _type = "Task" [ 619.106417] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.116593] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52366b33-1bf8-9dba-e73a-db36f3f4d10b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.219817] env[65385]: DEBUG nova.compute.manager [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 619.294060] env[65385]: DEBUG oslo_vmware.api [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Task: {'id': task-4453151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22717} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.294060] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 619.294060] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 619.294060] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.294060] env[65385]: INFO nova.compute.manager [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Took 1.15 seconds to destroy the instance on the hypervisor. [ 619.295075] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 619.295075] env[65385]: DEBUG nova.compute.manager [-] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 619.295075] env[65385]: DEBUG nova.network.neutron [-] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 619.295075] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.295075] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.295075] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.313015] env[65385]: WARNING neutronclient.v2_0.client [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.313675] env[65385]: WARNING openstack [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.315035] env[65385]: WARNING openstack [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.554018] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Releasing lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.554018] env[65385]: DEBUG nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Instance network_info: |[{"id": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "address": "fa:16:3e:a7:7e:98", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape37bf46a-30", "ovs_interfaceid": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 619.554407] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:7e:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e37bf46a-308d-44a5-b573-64c3b7aa5a73', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 619.560985] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Creating folder: Project (e94f2a868e2d4d31af54f2e5d8c479f8). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.564133] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79fc65a0-5b9d-4555-88d9-fe1451389e98 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.579830] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453145, 'name': ReconfigVM_Task, 'duration_secs': 1.796892} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.583765] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Reconfigured VM instance instance-0000000c to attach disk [datastore2] 649b4e21-fc55-415c-a1f3-ec724397b874/649b4e21-fc55-415c-a1f3-ec724397b874.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 619.584648] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Created folder: Project (e94f2a868e2d4d31af54f2e5d8c479f8) in parent group-v870881. [ 619.584802] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Creating folder: Instances. Parent ref: group-v870922. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.585754] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-210b09e1-611a-45e6-a617-d75686feb35f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.587023] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a87c2f8-8e92-4052-9103-4cdf0eb946a0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.589217] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.597563] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 619.597563] env[65385]: value = "task-4453154" [ 619.597563] env[65385]: _type = "Task" [ 619.597563] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.607799] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Created folder: Instances in parent group-v870922. [ 619.608233] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 619.608885] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 619.616366] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c72f52b-5315-4292-9835-c853bd5fb6d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.631632] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453154, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.643965] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52366b33-1bf8-9dba-e73a-db36f3f4d10b, 'name': SearchDatastore_Task, 'duration_secs': 0.014027} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.645540] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.645900] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] f77d861a-ae66-44c3-901d-e499926d6f5b/f77d861a-ae66-44c3-901d-e499926d6f5b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 619.646531] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 619.646531] env[65385]: value = "task-4453155" [ 619.646531] env[65385]: _type = "Task" [ 619.646531] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.646531] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-798038cc-23da-41ad-b03a-7f28582e6268 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.660751] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453155, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.662329] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 619.662329] env[65385]: value = "task-4453156" [ 619.662329] env[65385]: _type = "Task" [ 619.662329] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.676547] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.742135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.848283] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d3f316-35da-49eb-946a-19c93f6fb1df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.864040] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00e102d-97ce-40ef-ace4-21c02aed9d2f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.902119] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94da3df0-5f65-463c-9314-b29680bb02da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.915325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426037dd-b3ca-4525-b27b-9be5ffc279eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.936789] env[65385]: DEBUG nova.compute.provider_tree [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.113813] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453154, 'name': Rename_Task, 'duration_secs': 0.197278} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.113813] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 620.113813] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-073a729a-d290-43b2-91c5-66ad305ca7cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.123777] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 620.123777] env[65385]: value = "task-4453157" [ 620.123777] env[65385]: _type = "Task" [ 620.123777] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.136559] env[65385]: DEBUG nova.network.neutron [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Successfully updated port: 99c91233-f17e-45f0-8852-fd19e18a8502 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 620.141606] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.169800] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453155, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.186296] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453156, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.188484] env[65385]: WARNING openstack [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.189110] env[65385]: WARNING openstack [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.323166] env[65385]: WARNING neutronclient.v2_0.client [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 620.324098] env[65385]: WARNING openstack [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.324493] env[65385]: WARNING openstack [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.444745] env[65385]: DEBUG nova.network.neutron [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Updated VIF entry in instance network info cache for port 7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 620.445436] env[65385]: DEBUG nova.network.neutron [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Updating instance_info_cache with network_info: [{"id": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "address": "fa:16:3e:5c:e2:96", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a4f4b0b-f7", "ovs_interfaceid": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 620.447448] env[65385]: DEBUG nova.scheduler.client.report [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 620.493710] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "75ba706f-2e72-4f84-b02f-db4381951e77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 620.493995] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "75ba706f-2e72-4f84-b02f-db4381951e77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.587958] env[65385]: DEBUG nova.network.neutron [-] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 620.626551] env[65385]: DEBUG nova.compute.manager [None req-c8779653-b14a-4cf9-b98b-679bd0e45f08 tempest-ServerDiagnosticsV248Test-718726699 tempest-ServerDiagnosticsV248Test-718726699-project-admin] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 620.631227] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec7fe8d-f882-4c9e-b39b-47a1bddad133 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.641477] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453157, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.643600] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "refresh_cache-5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.643752] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "refresh_cache-5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.643949] env[65385]: DEBUG nova.network.neutron [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 620.645825] env[65385]: INFO nova.compute.manager [None req-c8779653-b14a-4cf9-b98b-679bd0e45f08 tempest-ServerDiagnosticsV248Test-718726699 tempest-ServerDiagnosticsV248Test-718726699-project-admin] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Retrieving diagnostics [ 620.647027] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b2a9fd-3d33-4097-8137-d904ca6f4276 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.660890] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453155, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.706522] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589642} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.706857] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] f77d861a-ae66-44c3-901d-e499926d6f5b/f77d861a-ae66-44c3-901d-e499926d6f5b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 620.707124] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 620.707455] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84fa7608-dac0-4225-a28c-c1b264fc7051 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.716820] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 620.716820] env[65385]: value = "task-4453158" [ 620.716820] env[65385]: _type = "Task" [ 620.716820] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.727313] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453158, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.951017] env[65385]: DEBUG oslo_concurrency.lockutils [req-9c2e60f6-9955-4599-b4ba-72e9055edbfa req-6abe2794-235d-469d-b29b-d71a599a6f00 service nova] Releasing lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.952055] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 620.952660] env[65385]: DEBUG nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 620.956134] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.623s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.956452] env[65385]: DEBUG nova.objects.instance [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lazy-loading 'resources' on Instance uuid b72885c3-5146-42a5-82e8-444cfd89413d {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 620.996475] env[65385]: DEBUG nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 621.091749] env[65385]: INFO nova.compute.manager [-] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Took 1.80 seconds to deallocate network for instance. [ 621.139690] env[65385]: DEBUG oslo_vmware.api [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453157, 'name': PowerOnVM_Task, 'duration_secs': 0.823083} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.139955] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 621.140433] env[65385]: INFO nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Took 10.86 seconds to spawn the instance on the hypervisor. [ 621.140594] env[65385]: DEBUG nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 621.141397] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5efab5-078c-4b11-96de-1940e0aba1d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.147897] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.148324] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.175992] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453155, 'name': CreateVM_Task, 'duration_secs': 1.472966} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.176218] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 621.176715] env[65385]: WARNING neutronclient.v2_0.client [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.177080] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.177223] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.177953] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 621.178563] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06e07ff4-5826-4d6b-b0ad-01da2c740b04 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.184934] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 621.184934] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525b7cb6-93c5-e918-8b6f-42f89da911ca" [ 621.184934] env[65385]: _type = "Task" [ 621.184934] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.194481] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525b7cb6-93c5-e918-8b6f-42f89da911ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.223716] env[65385]: DEBUG nova.network.neutron [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 621.230450] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453158, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072411} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.230450] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.233672] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd39aa5f-40ed-4e32-a14b-e058fe181c2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.259620] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] f77d861a-ae66-44c3-901d-e499926d6f5b/f77d861a-ae66-44c3-901d-e499926d6f5b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.262018] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.263075] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.270717] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2a1b428-e47a-4b7a-9544-3de03791be75 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.298099] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 621.298099] env[65385]: value = "task-4453159" [ 621.298099] env[65385]: _type = "Task" [ 621.298099] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.309178] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453159, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.364027] env[65385]: WARNING neutronclient.v2_0.client [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.364027] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.364027] env[65385]: WARNING openstack [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.460148] env[65385]: DEBUG nova.compute.utils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 621.461827] env[65385]: DEBUG nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 621.463309] env[65385]: DEBUG nova.network.neutron [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 621.463718] env[65385]: WARNING neutronclient.v2_0.client [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.465142] env[65385]: WARNING neutronclient.v2_0.client [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.465142] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.465427] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.479212] env[65385]: DEBUG nova.network.neutron [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Updating instance_info_cache with network_info: [{"id": "99c91233-f17e-45f0-8852-fd19e18a8502", "address": "fa:16:3e:3b:31:cf", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99c91233-f1", "ovs_interfaceid": "99c91233-f17e-45f0-8852-fd19e18a8502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 621.521271] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.527987] env[65385]: DEBUG nova.policy [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bbc524f6b2e64dd0af9d06525190299e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99f640b7d5594400bbd69c05713089d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 621.600045] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.677173] env[65385]: INFO nova.compute.manager [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Took 22.50 seconds to build instance. [ 621.699054] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525b7cb6-93c5-e918-8b6f-42f89da911ca, 'name': SearchDatastore_Task, 'duration_secs': 0.019431} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.699241] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 621.699501] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 621.699750] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.699994] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.700190] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 621.701212] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-390cf05e-bffe-4605-b4f0-96023119946b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.719264] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 621.719448] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 621.720408] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1b393aa-2e1f-4821-973b-699c699c51d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.731014] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 621.731014] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524a7ff0-a8a1-54d1-a503-1930337522a8" [ 621.731014] env[65385]: _type = "Task" [ 621.731014] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.740918] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524a7ff0-a8a1-54d1-a503-1930337522a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.811199] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453159, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.882455] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00a3ba2-ad9e-4d6d-a915-de9e55ccaa57 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.897519] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d5c08b-55f8-4892-a407-c81225fa2723 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.938019] env[65385]: DEBUG nova.network.neutron [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Successfully created port: 82373c5b-5d69-406c-9fca-ed283c2a5f88 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 621.941030] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e790b7c7-07d2-4f5d-bf76-c4340080cccc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.952014] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a126dd3-5b10-45aa-9e8c-baedf4753711 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.969731] env[65385]: DEBUG nova.compute.provider_tree [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.977296] env[65385]: DEBUG nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 621.981896] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "refresh_cache-5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 621.983767] env[65385]: DEBUG nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Instance network_info: |[{"id": "99c91233-f17e-45f0-8852-fd19e18a8502", "address": "fa:16:3e:3b:31:cf", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99c91233-f1", "ovs_interfaceid": "99c91233-f17e-45f0-8852-fd19e18a8502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 621.983965] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:31:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7514a465-f1a4-4a8b-b76b-726b1a9d7e2f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99c91233-f17e-45f0-8852-fd19e18a8502', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 621.990898] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 621.991484] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 621.991730] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17db3e37-1f99-470b-9e87-8e35827281d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.014200] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 622.014200] env[65385]: value = "task-4453160" [ 622.014200] env[65385]: _type = "Task" [ 622.014200] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.024411] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453160, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.180710] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11757cf1-3c59-4d90-b052-974cb7bd3388 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.015s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.245623] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524a7ff0-a8a1-54d1-a503-1930337522a8, 'name': SearchDatastore_Task, 'duration_secs': 0.020473} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.246717] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-445193e8-f048-4b5b-b359-1a6435db00f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.255373] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 622.255373] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eb80b2-0888-98f6-ec1c-99f531c06440" [ 622.255373] env[65385]: _type = "Task" [ 622.255373] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.265233] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eb80b2-0888-98f6-ec1c-99f531c06440, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.311185] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453159, 'name': ReconfigVM_Task, 'duration_secs': 0.638591} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.311785] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Reconfigured VM instance instance-0000000d to attach disk [datastore2] f77d861a-ae66-44c3-901d-e499926d6f5b/f77d861a-ae66-44c3-901d-e499926d6f5b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.312780] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af588c00-ca9d-4dde-8774-99ab6c4d274a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.323428] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 622.323428] env[65385]: value = "task-4453161" [ 622.323428] env[65385]: _type = "Task" [ 622.323428] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.336311] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453161, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.472754] env[65385]: DEBUG nova.scheduler.client.report [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 622.527926] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453160, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.768030] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eb80b2-0888-98f6-ec1c-99f531c06440, 'name': SearchDatastore_Task, 'duration_secs': 0.013745} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.768331] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.769935] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0115dde8-b0aa-4960-a452-ced6587bc567/0115dde8-b0aa-4960-a452-ced6587bc567.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 622.769935] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24cd938e-c1b5-4301-8b62-c67f28025433 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.776933] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 622.776933] env[65385]: value = "task-4453162" [ 622.776933] env[65385]: _type = "Task" [ 622.776933] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.787387] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453162, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.834686] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453161, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.978094] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.022s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.981882] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.658s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.984439] env[65385]: INFO nova.compute.claims [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.988883] env[65385]: DEBUG nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 623.006523] env[65385]: INFO nova.scheduler.client.report [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Deleted allocations for instance b72885c3-5146-42a5-82e8-444cfd89413d [ 623.025140] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 623.025140] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 623.025420] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 623.025753] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 623.026405] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 623.026405] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 623.026405] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.028141] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 623.028141] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 623.028141] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 623.028141] env[65385]: DEBUG nova.virt.hardware [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 623.028969] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80162710-2ad2-4882-b661-aa374c6b9380 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.035817] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453160, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.043639] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6be40b0-a9bb-48ed-bad1-7b44828c8899 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.290063] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453162, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.340599] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453161, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.522394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-505422ce-bbab-4e1a-ae4c-6a50facd8d7a tempest-ServerDiagnosticsTest-1955242197 tempest-ServerDiagnosticsTest-1955242197-project-member] Lock "b72885c3-5146-42a5-82e8-444cfd89413d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.091s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 623.529681] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453160, 'name': CreateVM_Task, 'duration_secs': 1.438981} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.529876] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 623.530359] env[65385]: WARNING neutronclient.v2_0.client [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 623.530732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.530875] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.531238] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 623.531529] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b82ec1d2-c79c-483e-8fd9-2da053f07336 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.537401] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 623.537401] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a14ebb-b51f-4c1b-511c-63a550c6e729" [ 623.537401] env[65385]: _type = "Task" [ 623.537401] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.549163] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a14ebb-b51f-4c1b-511c-63a550c6e729, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.791488] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453162, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.748194} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.791851] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0115dde8-b0aa-4960-a452-ced6587bc567/0115dde8-b0aa-4960-a452-ced6587bc567.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 623.792011] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 623.792322] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84e57817-a50f-4269-bc15-e0c7b269792f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.801106] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 623.801106] env[65385]: value = "task-4453163" [ 623.801106] env[65385]: _type = "Task" [ 623.801106] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.811301] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453163, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.838881] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453161, 'name': Rename_Task, 'duration_secs': 1.215692} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.838881] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 623.838881] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3090732f-ad58-4e56-b838-34d0e3b11311 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.845617] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 623.845617] env[65385]: value = "task-4453164" [ 623.845617] env[65385]: _type = "Task" [ 623.845617] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.859168] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453164, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.864674] env[65385]: DEBUG nova.network.neutron [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Successfully updated port: 82373c5b-5d69-406c-9fca-ed283c2a5f88 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 624.056812] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a14ebb-b51f-4c1b-511c-63a550c6e729, 'name': SearchDatastore_Task, 'duration_secs': 0.013687} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.058008] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.058303] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 624.058613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.058853] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.059085] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 624.059431] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-755792c6-dbfe-4bc3-9510-be7123563b41 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.083047] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 624.083276] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 624.084121] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6a8b9d9-7ec6-4a41-8442-af705f70e3d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.091688] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 624.091688] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521a94c3-b9c2-206a-1d83-60cc5729564d" [ 624.091688] env[65385]: _type = "Task" [ 624.091688] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.103533] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521a94c3-b9c2-206a-1d83-60cc5729564d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.314552] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453163, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077497} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.317312] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 624.319149] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4e2d62-b232-46e4-b448-02d685b8f80e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.350871] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] 0115dde8-b0aa-4960-a452-ced6587bc567/0115dde8-b0aa-4960-a452-ced6587bc567.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 624.354161] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4be526d-0b57-4043-9425-391a9aef396b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.379579] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "refresh_cache-60067529-1071-4295-b1c9-21523bf347d9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.379765] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "refresh_cache-60067529-1071-4295-b1c9-21523bf347d9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.379942] env[65385]: DEBUG nova.network.neutron [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 624.389678] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 624.389678] env[65385]: value = "task-4453165" [ 624.389678] env[65385]: _type = "Task" [ 624.389678] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.392999] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453164, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.406811] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.478259] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3ecec7-6909-4d20-a27c-08fdc4142857 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.489211] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1078f9-422c-400e-a5d7-08fdb47f5345 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.522996] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567e062b-72c6-488d-8df9-f624e3fc28a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.532378] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b44fa1d-0df7-42ee-907d-3fb85b98cf33 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.549131] env[65385]: DEBUG nova.compute.provider_tree [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.607314] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521a94c3-b9c2-206a-1d83-60cc5729564d, 'name': SearchDatastore_Task, 'duration_secs': 0.025664} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.611761] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5729ec5f-b223-4113-a974-83d5a85095cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.619364] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 624.619364] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523f5623-a400-58b8-669c-958d18e4bde7" [ 624.619364] env[65385]: _type = "Task" [ 624.619364] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.631785] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523f5623-a400-58b8-669c-958d18e4bde7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.644770] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Received event network-vif-plugged-34c764b8-f45c-4237-9998-f49b871bb7d4 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 624.645038] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Acquiring lock "f77d861a-ae66-44c3-901d-e499926d6f5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.645221] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.645380] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.645548] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] No waiting events found dispatching network-vif-plugged-34c764b8-f45c-4237-9998-f49b871bb7d4 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 624.645705] env[65385]: WARNING nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Received unexpected event network-vif-plugged-34c764b8-f45c-4237-9998-f49b871bb7d4 for instance with vm_state building and task_state spawning. [ 624.645866] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Received event network-changed-34c764b8-f45c-4237-9998-f49b871bb7d4 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 624.645991] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Refreshing instance network info cache due to event network-changed-34c764b8-f45c-4237-9998-f49b871bb7d4. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 624.646187] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Acquiring lock "refresh_cache-f77d861a-ae66-44c3-901d-e499926d6f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.646314] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Acquired lock "refresh_cache-f77d861a-ae66-44c3-901d-e499926d6f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.646506] env[65385]: DEBUG nova.network.neutron [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Refreshing network info cache for port 34c764b8-f45c-4237-9998-f49b871bb7d4 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 624.667834] env[65385]: DEBUG nova.compute.manager [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Received event network-vif-plugged-e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 624.668304] env[65385]: DEBUG oslo_concurrency.lockutils [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Acquiring lock "0115dde8-b0aa-4960-a452-ced6587bc567-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.668546] env[65385]: DEBUG oslo_concurrency.lockutils [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Lock "0115dde8-b0aa-4960-a452-ced6587bc567-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.668546] env[65385]: DEBUG oslo_concurrency.lockutils [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Lock "0115dde8-b0aa-4960-a452-ced6587bc567-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.668722] env[65385]: DEBUG nova.compute.manager [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] No waiting events found dispatching network-vif-plugged-e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 624.668875] env[65385]: WARNING nova.compute.manager [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Received unexpected event network-vif-plugged-e37bf46a-308d-44a5-b573-64c3b7aa5a73 for instance with vm_state building and task_state spawning. [ 624.669040] env[65385]: DEBUG nova.compute.manager [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Received event network-changed-e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 624.669189] env[65385]: DEBUG nova.compute.manager [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Refreshing instance network info cache due to event network-changed-e37bf46a-308d-44a5-b573-64c3b7aa5a73. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 624.669356] env[65385]: DEBUG oslo_concurrency.lockutils [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Acquiring lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.669482] env[65385]: DEBUG oslo_concurrency.lockutils [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Acquired lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.669634] env[65385]: DEBUG nova.network.neutron [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Refreshing network info cache for port e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 624.864805] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453164, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.886219] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.886629] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 624.903150] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453165, 'name': ReconfigVM_Task, 'duration_secs': 0.338573} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.903423] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Reconfigured VM instance instance-0000000e to attach disk [datastore2] 0115dde8-b0aa-4960-a452-ced6587bc567/0115dde8-b0aa-4960-a452-ced6587bc567.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 624.904092] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53c3895f-05dd-4cbd-bc86-29059d3c4c56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.911331] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 624.911331] env[65385]: value = "task-4453166" [ 624.911331] env[65385]: _type = "Task" [ 624.911331] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.911783] env[65385]: INFO nova.compute.manager [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Rebuilding instance [ 624.924185] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453166, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.983768] env[65385]: DEBUG nova.compute.manager [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 624.983768] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbde28ec-eea7-4066-b69c-f264391e3616 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.054299] env[65385]: DEBUG nova.scheduler.client.report [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 625.116572] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "ef235d8e-6193-4010-ac50-72cd275e965a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.117818] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "ef235d8e-6193-4010-ac50-72cd275e965a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.117818] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "ef235d8e-6193-4010-ac50-72cd275e965a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.118184] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "ef235d8e-6193-4010-ac50-72cd275e965a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.118392] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "ef235d8e-6193-4010-ac50-72cd275e965a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.124762] env[65385]: INFO nova.compute.manager [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Terminating instance [ 625.134900] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523f5623-a400-58b8-669c-958d18e4bde7, 'name': SearchDatastore_Task, 'duration_secs': 0.020569} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.135226] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.135478] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64/5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 625.135695] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb910895-15f7-4661-b442-78bf73ad9ea0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.144034] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 625.144034] env[65385]: value = "task-4453167" [ 625.144034] env[65385]: _type = "Task" [ 625.144034] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.149246] env[65385]: WARNING neutronclient.v2_0.client [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.149774] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.150159] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.163670] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453167, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.172642] env[65385]: WARNING neutronclient.v2_0.client [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.173488] env[65385]: WARNING openstack [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.174492] env[65385]: WARNING openstack [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.366022] env[65385]: DEBUG oslo_vmware.api [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453164, 'name': PowerOnVM_Task, 'duration_secs': 1.2336} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.366744] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 625.366744] env[65385]: INFO nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Took 11.99 seconds to spawn the instance on the hypervisor. [ 625.366744] env[65385]: DEBUG nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 625.367660] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a521203-10f8-45b1-acce-81c8524b5186 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.423555] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453166, 'name': Rename_Task, 'duration_secs': 0.317173} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.424256] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 625.424638] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef0407c3-953d-4877-b704-6ed46f5fccdc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.434432] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 625.434432] env[65385]: value = "task-4453168" [ 625.434432] env[65385]: _type = "Task" [ 625.434432] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.443640] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.551550] env[65385]: DEBUG nova.network.neutron [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 625.561368] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.561368] env[65385]: DEBUG nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 625.564077] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.045s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.566095] env[65385]: INFO nova.compute.claims [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.631284] env[65385]: DEBUG nova.compute.manager [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 625.631284] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.632111] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb7d16f-b57b-49c0-9afb-2cb3e2c62d87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.642604] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 625.642989] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-557fc183-9fd6-48ec-9b08-c2bc4f507e3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.657489] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453167, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.659772] env[65385]: DEBUG oslo_vmware.api [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 625.659772] env[65385]: value = "task-4453169" [ 625.659772] env[65385]: _type = "Task" [ 625.659772] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.672871] env[65385]: DEBUG oslo_vmware.api [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.897225] env[65385]: INFO nova.compute.manager [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Took 24.23 seconds to build instance. [ 625.952069] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453168, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.000695] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 626.001054] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9cdf9cf-06f1-4e3f-b359-f7400b0c57c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.012107] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 626.012107] env[65385]: value = "task-4453170" [ 626.012107] env[65385]: _type = "Task" [ 626.012107] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.030665] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.072682] env[65385]: DEBUG nova.compute.utils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 626.076978] env[65385]: DEBUG nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 626.077240] env[65385]: DEBUG nova.network.neutron [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 626.077572] env[65385]: WARNING neutronclient.v2_0.client [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.077870] env[65385]: WARNING neutronclient.v2_0.client [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.078478] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.078827] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.157116] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453167, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727723} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.157423] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64/5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 626.157689] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 626.161379] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e2bfaa2-578e-4cb3-b161-ae06a1fcbf94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.180952] env[65385]: DEBUG oslo_vmware.api [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453169, 'name': PowerOffVM_Task, 'duration_secs': 0.218214} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.183205] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 626.183409] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 626.183602] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 626.183602] env[65385]: value = "task-4453171" [ 626.183602] env[65385]: _type = "Task" [ 626.183602] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.183828] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83b16aa4-f20c-4bc4-8357-33e6b6024af5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.197820] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453171, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.274799] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 626.275044] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 626.275236] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Deleting the datastore file [datastore1] ef235d8e-6193-4010-ac50-72cd275e965a {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 626.275484] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-977c13f9-ece7-4015-93ba-7aafb1183f5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.283822] env[65385]: DEBUG oslo_vmware.api [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for the task: (returnval){ [ 626.283822] env[65385]: value = "task-4453173" [ 626.283822] env[65385]: _type = "Task" [ 626.283822] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.297084] env[65385]: DEBUG oslo_vmware.api [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453173, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.399494] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2e1ae783-a308-4c24-9c84-1ca41aa49a98 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.737s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 626.448518] env[65385]: DEBUG oslo_vmware.api [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453168, 'name': PowerOnVM_Task, 'duration_secs': 0.568696} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.449288] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 626.450053] env[65385]: INFO nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Took 10.47 seconds to spawn the instance on the hypervisor. [ 626.450053] env[65385]: DEBUG nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 626.451158] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31d3046-e9f2-46d0-ac01-8f02237e155c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.500749] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.501158] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.524009] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453170, 'name': PowerOffVM_Task, 'duration_secs': 0.371162} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.524009] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 626.524009] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.524894] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d806f89-b6be-49b4-b71d-7b883706ef21 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.534291] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 626.536629] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47ba5382-a5d6-40c4-a834-193b07dd4f91 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.565302] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 626.565302] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 626.565302] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Deleting the datastore file [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 626.565302] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e21337a-517c-451f-83df-8428ab615c5a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.574147] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 626.574147] env[65385]: value = "task-4453175" [ 626.574147] env[65385]: _type = "Task" [ 626.574147] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.578475] env[65385]: DEBUG nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 626.593187] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453175, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.673783] env[65385]: DEBUG nova.policy [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7cda854e4ee54615b5510bb670939a40', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f858126fa23d43cbafd1b677206f28ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 626.699073] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453171, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082224} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.699442] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 626.702776] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c557c39b-4b81-4a61-85b7-195dfecf4ee3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.728481] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64/5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 626.732144] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a0d4c87-d9d6-44d6-81f8-aea5f9c06d19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.757851] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 626.757851] env[65385]: value = "task-4453176" [ 626.757851] env[65385]: _type = "Task" [ 626.757851] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.769963] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453176, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.800032] env[65385]: DEBUG oslo_vmware.api [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453173, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.972969] env[65385]: INFO nova.compute.manager [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Took 25.17 seconds to build instance. [ 627.023919] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c3229e-6dd3-4b15-9525-fd1fe6f547ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.033413] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50dfaa0-2a32-46f1-8a3a-99157712f6c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.067873] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1e805b-0fb1-48ff-841e-19a55f43d695 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.082846] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72d0dfc-c0bf-41fe-a691-cb8ac9ac18c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.107300] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453175, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.477096} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.107833] env[65385]: DEBUG nova.compute.provider_tree [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.109606] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 627.109894] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 627.110115] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.271889] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453176, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.300795] env[65385]: DEBUG oslo_vmware.api [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Task: {'id': task-4453173, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.658934} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.301082] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 627.301318] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 627.301449] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.301949] env[65385]: INFO nova.compute.manager [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Took 1.67 seconds to destroy the instance on the hypervisor. [ 627.301949] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 627.302324] env[65385]: DEBUG nova.compute.manager [-] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 627.302324] env[65385]: DEBUG nova.network.neutron [-] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 627.302589] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.303260] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.303560] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.478395] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fc3f0ad8-e52c-4eb0-b8c8-802e1c509630 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "0115dde8-b0aa-4960-a452-ced6587bc567" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.688s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 627.596196] env[65385]: DEBUG nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 627.611942] env[65385]: DEBUG nova.scheduler.client.report [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 627.636823] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 627.639727] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 627.642227] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 627.642227] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 627.642227] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 627.642227] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 627.642227] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.642583] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 627.642583] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 627.642583] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 627.642583] env[65385]: DEBUG nova.virt.hardware [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 627.644685] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87aed8c2-5b5f-48b8-bc5b-294e967633a1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.655414] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3347c6d6-b5cc-4f25-a080-75bee4ec2d19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.773935] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453176, 'name': ReconfigVM_Task, 'duration_secs': 0.738617} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.774585] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64/5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 627.775637] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3b290ae-e8a8-4ba4-a431-f84e98c7cf45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.784632] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 627.784632] env[65385]: value = "task-4453177" [ 627.784632] env[65385]: _type = "Task" [ 627.784632] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.797086] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453177, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.808061] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.808491] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.826377] env[65385]: WARNING openstack [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.826451] env[65385]: WARNING openstack [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.904679] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.908498] env[65385]: WARNING neutronclient.v2_0.client [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.909207] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.909441] env[65385]: WARNING openstack [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.120193] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.121160] env[65385]: DEBUG nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 628.125039] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.240s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 628.125283] env[65385]: DEBUG nova.objects.instance [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lazy-loading 'resources' on Instance uuid 6e1eeff2-1587-44c6-a154-f73a2d0f8061 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 628.165696] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 628.166056] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 628.166411] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 628.166513] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 628.166625] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 628.166747] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 628.166986] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.167513] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 628.167600] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 628.167767] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 628.167953] env[65385]: DEBUG nova.virt.hardware [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 628.169485] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c11f0e-bd76-4753-95cb-e3d53c906efe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.178749] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e9fd94-0cf7-4c02-94f8-d7a16f696adf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.200756] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 628.206234] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 628.206519] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 628.206732] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f11e0168-4f30-4486-b9cc-a330f63051e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.232820] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 628.232820] env[65385]: value = "task-4453178" [ 628.232820] env[65385]: _type = "Task" [ 628.232820] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.243514] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453178, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.304060] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453177, 'name': Rename_Task, 'duration_secs': 0.413273} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.304060] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 628.304592] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e7988f3-7f23-42ac-b618-8ca2380e2b20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.317563] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 628.317563] env[65385]: value = "task-4453179" [ 628.317563] env[65385]: _type = "Task" [ 628.317563] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.336865] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453179, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.398975] env[65385]: DEBUG nova.network.neutron [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Successfully created port: 97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 628.416318] env[65385]: WARNING neutronclient.v2_0.client [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.416978] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.417723] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.430038] env[65385]: WARNING neutronclient.v2_0.client [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.431016] env[65385]: WARNING openstack [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.431603] env[65385]: WARNING openstack [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.606071] env[65385]: DEBUG nova.network.neutron [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Updating instance_info_cache with network_info: [{"id": "82373c5b-5d69-406c-9fca-ed283c2a5f88", "address": "fa:16:3e:4c:47:90", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82373c5b-5d", "ovs_interfaceid": "82373c5b-5d69-406c-9fca-ed283c2a5f88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 628.628869] env[65385]: DEBUG nova.compute.utils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 628.630526] env[65385]: DEBUG nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 628.630721] env[65385]: DEBUG nova.network.neutron [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 628.631090] env[65385]: WARNING neutronclient.v2_0.client [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.631589] env[65385]: WARNING neutronclient.v2_0.client [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.632224] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.633033] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.760036] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453178, 'name': CreateVM_Task, 'duration_secs': 0.365731} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.760036] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 628.760036] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.760036] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.760036] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 628.760036] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2999567-6ced-43a5-8286-298f76e582f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.766129] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 628.766129] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52235e56-7170-6475-0064-2e70f63de76b" [ 628.766129] env[65385]: _type = "Task" [ 628.766129] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.784817] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52235e56-7170-6475-0064-2e70f63de76b, 'name': SearchDatastore_Task, 'duration_secs': 0.01182} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.789972] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.790252] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 628.790478] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.790618] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.790793] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 628.794015] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44871724-5aab-4862-8872-292c5feb45c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.807286] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 628.807513] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 628.808330] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ca9db53-cd8e-483e-bf8f-c9ac778a5f95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.825917] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 628.825917] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524055b4-b112-b09b-33ad-3ab862af008b" [ 628.825917] env[65385]: _type = "Task" [ 628.825917] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.842896] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453179, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.850686] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524055b4-b112-b09b-33ad-3ab862af008b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.886579] env[65385]: DEBUG nova.policy [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53765f0467d14fc3bc946cc893f75280', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '112402342b0e481f83be7e17c9f739e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 629.024759] env[65385]: DEBUG nova.network.neutron [-] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 629.036621] env[65385]: DEBUG nova.network.neutron [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Updated VIF entry in instance network info cache for port 34c764b8-f45c-4237-9998-f49b871bb7d4. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 629.037055] env[65385]: DEBUG nova.network.neutron [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Updating instance_info_cache with network_info: [{"id": "34c764b8-f45c-4237-9998-f49b871bb7d4", "address": "fa:16:3e:ca:cd:bf", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34c764b8-f4", "ovs_interfaceid": "34c764b8-f45c-4237-9998-f49b871bb7d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 629.049700] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fee5f45-f9ce-4eb2-af17-1db729164e2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.059473] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc42a2e4-58be-4515-ac6f-9e23fe45bb45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.094077] env[65385]: DEBUG nova.network.neutron [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updated VIF entry in instance network info cache for port e37bf46a-308d-44a5-b573-64c3b7aa5a73. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 629.094466] env[65385]: DEBUG nova.network.neutron [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updating instance_info_cache with network_info: [{"id": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "address": "fa:16:3e:a7:7e:98", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape37bf46a-30", "ovs_interfaceid": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 629.096124] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69e4d4a-2154-4f7f-a949-e32ba1837796 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.106522] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429debc5-5d9c-4726-9282-7cc7c31b5fd7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.113310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "refresh_cache-60067529-1071-4295-b1c9-21523bf347d9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.113905] env[65385]: DEBUG nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Instance network_info: |[{"id": "82373c5b-5d69-406c-9fca-ed283c2a5f88", "address": "fa:16:3e:4c:47:90", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82373c5b-5d", "ovs_interfaceid": "82373c5b-5d69-406c-9fca-ed283c2a5f88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 629.114125] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:47:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b89fd3b-0470-40c9-bb5b-d52c76c030e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82373c5b-5d69-406c-9fca-ed283c2a5f88', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 629.121535] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Creating folder: Project (99f640b7d5594400bbd69c05713089d3). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 629.122715] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-252caa61-a2d5-4888-8de0-65790cc1b115 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.133697] env[65385]: DEBUG nova.compute.provider_tree [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.136343] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Created folder: Project (99f640b7d5594400bbd69c05713089d3) in parent group-v870881. [ 629.136561] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Creating folder: Instances. Parent ref: group-v870927. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 629.137019] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48986954-3ebf-4e6f-be89-8c5e101e6ee0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.145539] env[65385]: DEBUG nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 629.149610] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Created folder: Instances in parent group-v870927. [ 629.149852] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 629.150332] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 629.150470] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a992254-f879-4e91-b681-92851cd57352 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.171861] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 629.171861] env[65385]: value = "task-4453182" [ 629.171861] env[65385]: _type = "Task" [ 629.171861] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.181666] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453182, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.335931] env[65385]: DEBUG oslo_vmware.api [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453179, 'name': PowerOnVM_Task, 'duration_secs': 0.562368} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.336666] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 629.336908] env[65385]: INFO nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Took 11.37 seconds to spawn the instance on the hypervisor. [ 629.337148] env[65385]: DEBUG nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 629.338029] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3eb1f4-8d9e-463f-9dfd-f79a919d77e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.345510] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524055b4-b112-b09b-33ad-3ab862af008b, 'name': SearchDatastore_Task, 'duration_secs': 0.032315} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.347237] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-147be701-8b88-4e4a-a3b4-d6c58dc7cf90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.361015] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 629.361015] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5246686b-fc0a-192d-f39b-bfde05e4b5ae" [ 629.361015] env[65385]: _type = "Task" [ 629.361015] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.374591] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5246686b-fc0a-192d-f39b-bfde05e4b5ae, 'name': SearchDatastore_Task, 'duration_secs': 0.01066} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.374591] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.374591] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 629.374591] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-865cdd31-fb7a-483e-aab7-95f01cc34705 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.389062] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 629.389062] env[65385]: value = "task-4453183" [ 629.389062] env[65385]: _type = "Task" [ 629.389062] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.397392] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.466201] env[65385]: DEBUG nova.network.neutron [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Successfully created port: 5f60e5e9-2860-4364-99c0-3c1433e4bddb {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 629.528634] env[65385]: INFO nova.compute.manager [-] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Took 2.23 seconds to deallocate network for instance. [ 629.541398] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Releasing lock "refresh_cache-f77d861a-ae66-44c3-901d-e499926d6f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.541765] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Received event network-vif-deleted-6cd369b2-83d9-43c0-8ba1-89eb0dd6f802 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 629.542032] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Received event network-vif-plugged-99c91233-f17e-45f0-8852-fd19e18a8502 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 629.542233] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Acquiring lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 629.543025] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 629.543025] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.543840] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] No waiting events found dispatching network-vif-plugged-99c91233-f17e-45f0-8852-fd19e18a8502 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 629.543840] env[65385]: WARNING nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Received unexpected event network-vif-plugged-99c91233-f17e-45f0-8852-fd19e18a8502 for instance with vm_state building and task_state spawning. [ 629.543840] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Received event network-changed-99c91233-f17e-45f0-8852-fd19e18a8502 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 629.543840] env[65385]: DEBUG nova.compute.manager [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Refreshing instance network info cache due to event network-changed-99c91233-f17e-45f0-8852-fd19e18a8502. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 629.543840] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Acquiring lock "refresh_cache-5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.544149] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Acquired lock "refresh_cache-5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.544904] env[65385]: DEBUG nova.network.neutron [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Refreshing network info cache for port 99c91233-f17e-45f0-8852-fd19e18a8502 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 629.599673] env[65385]: DEBUG oslo_concurrency.lockutils [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] Releasing lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.599947] env[65385]: DEBUG nova.compute.manager [req-5308b6a2-19b0-4870-9e9d-2a09736bff97 req-719a3c2f-f62d-43a3-b249-97f2ba72eed2 service nova] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Received event network-vif-deleted-87cd4460-4318-4258-a7b7-e1946f8dc8b8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 629.638820] env[65385]: DEBUG nova.scheduler.client.report [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.688357] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453182, 'name': CreateVM_Task, 'duration_secs': 0.414369} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.689685] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 629.689685] env[65385]: WARNING neutronclient.v2_0.client [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.689685] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.689814] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.690730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 629.690730] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee854d97-37af-4d1f-98e3-9d44f908d798 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.697737] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 629.697737] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520ef7a9-03b7-0898-4a3e-7a04b003938c" [ 629.697737] env[65385]: _type = "Task" [ 629.697737] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.715216] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520ef7a9-03b7-0898-4a3e-7a04b003938c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.871653] env[65385]: INFO nova.compute.manager [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Took 25.35 seconds to build instance. [ 629.900534] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453183, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.036735] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.048773] env[65385]: WARNING neutronclient.v2_0.client [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.049617] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.049979] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.144422] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.019s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.150038] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.136s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 630.150038] env[65385]: INFO nova.compute.claims [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.158030] env[65385]: DEBUG nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 630.176959] env[65385]: INFO nova.scheduler.client.report [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Deleted allocations for instance 6e1eeff2-1587-44c6-a154-f73a2d0f8061 [ 630.198737] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 630.200374] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 630.200374] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 630.200374] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 630.200374] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 630.200608] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 630.201066] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.201288] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 630.201448] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 630.201942] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 630.202025] env[65385]: DEBUG nova.virt.hardware [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 630.203925] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7615429f-3c53-4881-91cc-73ff3e0a96eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.224521] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a72db81-62d8-47d3-8ba2-2fb85633420a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.230387] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520ef7a9-03b7-0898-4a3e-7a04b003938c, 'name': SearchDatastore_Task, 'duration_secs': 0.072314} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.231635] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.234402] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 630.234402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.234402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.234402] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 630.234402] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-141e4979-09ae-4350-a22c-0858a7cb2482 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.247539] env[65385]: DEBUG nova.network.neutron [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Successfully updated port: 97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 630.260209] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 630.260474] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 630.261405] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7b41f8f-e907-4a90-880d-029c1ce1874e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.270467] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 630.270467] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5299a176-8b61-5761-3fcc-75c01b01e373" [ 630.270467] env[65385]: _type = "Task" [ 630.270467] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.283968] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5299a176-8b61-5761-3fcc-75c01b01e373, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.377989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f031e6f8-7a8b-41f0-988e-45eae91a9cf0 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.868s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.398843] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453183, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56534} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.399126] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 630.399335] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 630.399593] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09edc811-6e24-4536-b26b-dfe56d908400 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.408069] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 630.408069] env[65385]: value = "task-4453184" [ 630.408069] env[65385]: _type = "Task" [ 630.408069] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.418239] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.624190] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.624190] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.688117] env[65385]: DEBUG oslo_concurrency.lockutils [None req-718366fd-5498-4e09-857b-9e5fac55dc1e tempest-ServerDiagnosticsNegativeTest-266379322 tempest-ServerDiagnosticsNegativeTest-266379322-project-member] Lock "6e1eeff2-1587-44c6-a154-f73a2d0f8061" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.330s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.751013] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.751180] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquired lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.751365] env[65385]: DEBUG nova.network.neutron [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 630.777462] env[65385]: WARNING neutronclient.v2_0.client [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.778107] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.778743] env[65385]: WARNING openstack [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.799044] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5299a176-8b61-5761-3fcc-75c01b01e373, 'name': SearchDatastore_Task, 'duration_secs': 0.018563} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.800340] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21fce20b-f27e-464c-b891-cd94ce51e38c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.810903] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 630.810903] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52546a87-01e9-05b2-4541-c0c0b2abff23" [ 630.810903] env[65385]: _type = "Task" [ 630.810903] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.827740] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52546a87-01e9-05b2-4541-c0c0b2abff23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.926293] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074711} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.926584] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 630.928614] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10440580-1745-483f-bc9b-2dd9e4082e3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.955537] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 630.957981] env[65385]: DEBUG nova.network.neutron [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Updated VIF entry in instance network info cache for port 99c91233-f17e-45f0-8852-fd19e18a8502. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 630.958325] env[65385]: DEBUG nova.network.neutron [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Updating instance_info_cache with network_info: [{"id": "99c91233-f17e-45f0-8852-fd19e18a8502", "address": "fa:16:3e:3b:31:cf", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99c91233-f1", "ovs_interfaceid": "99c91233-f17e-45f0-8852-fd19e18a8502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 630.959532] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e684be2-dc11-4df0-88d3-ceff70bf3e02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.982699] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddea1a88-ccc8-4ada-bf17-7e697b5b282a req-ccf47fdd-0a79-4a39-92b5-8267987dca54 service nova] Releasing lock "refresh_cache-5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.987757] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 630.987757] env[65385]: value = "task-4453185" [ 630.987757] env[65385]: _type = "Task" [ 630.987757] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.998088] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.235646] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "69fdd5df-a9f2-486f-8a79-87c034366083" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 631.235646] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "69fdd5df-a9f2-486f-8a79-87c034366083" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 631.254098] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.254856] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.321932] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52546a87-01e9-05b2-4541-c0c0b2abff23, 'name': SearchDatastore_Task, 'duration_secs': 0.030032} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.325208] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.325573] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 60067529-1071-4295-b1c9-21523bf347d9/60067529-1071-4295-b1c9-21523bf347d9.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 631.326302] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc64b5c0-92d2-438f-a39d-e0bdab92b697 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.336512] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 631.336512] env[65385]: value = "task-4453186" [ 631.336512] env[65385]: _type = "Task" [ 631.336512] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.346609] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453186, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.391488] env[65385]: DEBUG nova.network.neutron [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Successfully updated port: 5f60e5e9-2860-4364-99c0-3c1433e4bddb {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 631.468292] env[65385]: DEBUG nova.network.neutron [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 631.500572] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.500762] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.517207] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453185, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.565854] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc098e6-3821-44c6-ba56-ab99ab28ed95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.578730] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047387d8-b38c-4e8b-9eba-c0c16e538553 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.620014] env[65385]: WARNING neutronclient.v2_0.client [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.620680] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.621091] env[65385]: WARNING openstack [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.629941] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce304d0-dc66-4072-a2b8-6c3c5d6a205d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.640688] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd5b1c4-953a-4e24-b745-dc4de8774a1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.661091] env[65385]: DEBUG nova.compute.provider_tree [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.709064] env[65385]: DEBUG nova.compute.manager [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Received event network-changed-6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 631.709064] env[65385]: DEBUG nova.compute.manager [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Refreshing instance network info cache due to event network-changed-6b5cfde5-8f86-4000-90b9-5c6140ea1606. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 631.709064] env[65385]: DEBUG oslo_concurrency.lockutils [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Acquiring lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.709265] env[65385]: DEBUG oslo_concurrency.lockutils [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Acquired lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.709770] env[65385]: DEBUG nova.network.neutron [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Refreshing network info cache for port 6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 631.735160] env[65385]: DEBUG nova.network.neutron [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Updating instance_info_cache with network_info: [{"id": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "address": "fa:16:3e:3a:2c:78", "network": {"id": "a2c91019-c13d-4581-aa0b-05553afc8f57", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1163204339-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f858126fa23d43cbafd1b677206f28ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97c1d2c0-12", "ovs_interfaceid": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 631.740153] env[65385]: DEBUG nova.compute.manager [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Received event network-changed-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 631.741170] env[65385]: DEBUG nova.compute.manager [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Refreshing instance network info cache due to event network-changed-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 631.741170] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Acquiring lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.741170] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Acquired lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.741170] env[65385]: DEBUG nova.network.neutron [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Refreshing network info cache for port f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 631.743748] env[65385]: DEBUG nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 631.849478] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453186, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496106} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.849734] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 60067529-1071-4295-b1c9-21523bf347d9/60067529-1071-4295-b1c9-21523bf347d9.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 631.849920] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 631.850153] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bcda58c8-8229-44e9-92d1-13a661f8fe0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.862043] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 631.862043] env[65385]: value = "task-4453187" [ 631.862043] env[65385]: _type = "Task" [ 631.862043] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.877173] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453187, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.897107] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "refresh_cache-1bc94ee7-d4f9-48c3-97f1-9e662eb50582" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.897107] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "refresh_cache-1bc94ee7-d4f9-48c3-97f1-9e662eb50582" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.897107] env[65385]: DEBUG nova.network.neutron [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 632.002577] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453185, 'name': ReconfigVM_Task, 'duration_secs': 0.620098} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.002857] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Reconfigured VM instance instance-00000005 to attach disk [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790/065659ae-2163-4d6b-a905-fc99d12b6790.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 632.003551] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-019ba6d7-01d4-4e6b-90a7-8932e560734a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.011732] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 632.011732] env[65385]: value = "task-4453188" [ 632.011732] env[65385]: _type = "Task" [ 632.011732] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.023121] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453188, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.165342] env[65385]: DEBUG nova.scheduler.client.report [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 632.213787] env[65385]: WARNING neutronclient.v2_0.client [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.214747] env[65385]: WARNING openstack [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.215347] env[65385]: WARNING openstack [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.245474] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Releasing lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.245999] env[65385]: DEBUG nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Instance network_info: |[{"id": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "address": "fa:16:3e:3a:2c:78", "network": {"id": "a2c91019-c13d-4581-aa0b-05553afc8f57", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1163204339-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f858126fa23d43cbafd1b677206f28ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97c1d2c0-12", "ovs_interfaceid": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 632.246121] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:2c:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97c1d2c0-12b5-428b-85b4-81b8305d87f9', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 632.257970] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Creating folder: Project (f858126fa23d43cbafd1b677206f28ac). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.258992] env[65385]: WARNING neutronclient.v2_0.client [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.260268] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.260528] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.272336] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77c9503f-a83a-481c-80bc-c2fdcb30e7cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.290046] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Created folder: Project (f858126fa23d43cbafd1b677206f28ac) in parent group-v870881. [ 632.290269] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Creating folder: Instances. Parent ref: group-v870930. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.290655] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e26e7327-2ef7-4dfb-97a4-0f274b00163c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.297922] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.303336] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Created folder: Instances in parent group-v870930. [ 632.303861] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 632.303861] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 632.304143] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5af2c5bf-580b-4b08-812c-759e3a126764 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.327100] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 632.327100] env[65385]: value = "task-4453191" [ 632.327100] env[65385]: _type = "Task" [ 632.327100] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.337971] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453191, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.373068] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453187, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070129} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.373372] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 632.374145] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d685f7-fe3e-44b3-9258-a4a5284feb00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.400454] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] 60067529-1071-4295-b1c9-21523bf347d9/60067529-1071-4295-b1c9-21523bf347d9.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 632.400968] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb3e4363-b386-42b4-8911-cef8f6ee90e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.416255] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.416637] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.430375] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 632.430375] env[65385]: value = "task-4453192" [ 632.430375] env[65385]: _type = "Task" [ 632.430375] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.439702] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453192, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.523978] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453188, 'name': Rename_Task, 'duration_secs': 0.260366} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.523978] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 632.523978] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22bd75ee-d33f-4101-9430-ed906086af62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.532480] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Waiting for the task: (returnval){ [ 632.532480] env[65385]: value = "task-4453193" [ 632.532480] env[65385]: _type = "Task" [ 632.532480] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.543210] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.670594] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.671225] env[65385]: DEBUG nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 632.674371] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.894s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.676058] env[65385]: INFO nova.compute.claims [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.811012] env[65385]: DEBUG nova.network.neutron [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 632.838186] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453191, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.874853] env[65385]: WARNING openstack [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.875431] env[65385]: WARNING openstack [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.946267] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453192, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.979637] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.980065] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.995827] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.996291] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.046279] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453193, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.168092] env[65385]: WARNING neutronclient.v2_0.client [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.168776] env[65385]: WARNING openstack [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.171726] env[65385]: WARNING openstack [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.183213] env[65385]: DEBUG nova.compute.utils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 633.189932] env[65385]: DEBUG nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 633.189932] env[65385]: DEBUG nova.network.neutron [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 633.189932] env[65385]: WARNING neutronclient.v2_0.client [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.190204] env[65385]: WARNING neutronclient.v2_0.client [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.190609] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.191253] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.199497] env[65385]: WARNING neutronclient.v2_0.client [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.201041] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.201041] env[65385]: WARNING openstack [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.253285] env[65385]: WARNING neutronclient.v2_0.client [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.254176] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.254802] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.341010] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453191, 'name': CreateVM_Task, 'duration_secs': 0.936384} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.341247] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 633.342391] env[65385]: WARNING neutronclient.v2_0.client [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.342872] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.343117] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.343559] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 633.343888] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01739bb4-9ad0-4bea-811e-85db13ec40e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.351586] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 633.351586] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52128ad5-4e6a-2c4b-eb79-afa1714f2542" [ 633.351586] env[65385]: _type = "Task" [ 633.351586] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.361157] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52128ad5-4e6a-2c4b-eb79-afa1714f2542, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.403262] env[65385]: DEBUG nova.policy [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27a258a335284ab3b73a6c55dfde1d6c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2de8098800694299aae5aa2f59f405bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 633.443650] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453192, 'name': ReconfigVM_Task, 'duration_secs': 0.83872} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.443957] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Reconfigured VM instance instance-00000010 to attach disk [datastore2] 60067529-1071-4295-b1c9-21523bf347d9/60067529-1071-4295-b1c9-21523bf347d9.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.444606] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c2654a54-861f-41d6-a27e-2364fe1c70f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.453545] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 633.453545] env[65385]: value = "task-4453194" [ 633.453545] env[65385]: _type = "Task" [ 633.453545] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.462834] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453194, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.546429] env[65385]: DEBUG oslo_vmware.api [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Task: {'id': task-4453193, 'name': PowerOnVM_Task, 'duration_secs': 0.835124} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.546429] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 633.546429] env[65385]: DEBUG nova.compute.manager [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 633.546970] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebacfd5a-1c4b-4f96-8ddc-7e09d6fe8363 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.569067] env[65385]: DEBUG nova.network.neutron [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Updated VIF entry in instance network info cache for port 6b5cfde5-8f86-4000-90b9-5c6140ea1606. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 633.569342] env[65385]: DEBUG nova.network.neutron [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Updating instance_info_cache with network_info: [{"id": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "address": "fa:16:3e:77:44:52", "network": {"id": "8919702d-2cd0-4d13-8f2f-31653142115a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-363641611-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a894ba2e7f7c42d091184408d0b84343", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5cfde5-8f", "ovs_interfaceid": "6b5cfde5-8f86-4000-90b9-5c6140ea1606", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 633.689968] env[65385]: DEBUG nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 633.722242] env[65385]: DEBUG nova.network.neutron [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Updating instance_info_cache with network_info: [{"id": "5f60e5e9-2860-4364-99c0-3c1433e4bddb", "address": "fa:16:3e:8e:64:4e", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f60e5e9-28", "ovs_interfaceid": "5f60e5e9-2860-4364-99c0-3c1433e4bddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 633.813681] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "de8632e0-bf38-440d-b6a3-895efeef122d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.814101] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "de8632e0-bf38-440d-b6a3-895efeef122d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.868406] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52128ad5-4e6a-2c4b-eb79-afa1714f2542, 'name': SearchDatastore_Task, 'duration_secs': 0.016887} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.869399] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.869544] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 633.870167] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.870167] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.870167] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 633.873550] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cef03c8c-a069-4126-b14b-46ace3943447 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.884533] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 633.884751] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 633.885495] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e58f50e-7d97-435d-aa98-aa3a9986ccc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.895773] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 633.895773] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52194848-71af-ae06-2a7c-925b3a268f7c" [ 633.895773] env[65385]: _type = "Task" [ 633.895773] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.905946] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52194848-71af-ae06-2a7c-925b3a268f7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.908260] env[65385]: DEBUG nova.network.neutron [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Updated VIF entry in instance network info cache for port f5dabab4-c5af-42f4-bfdc-eec4330bdb0d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 633.908588] env[65385]: DEBUG nova.network.neutron [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Updating instance_info_cache with network_info: [{"id": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "address": "fa:16:3e:0e:a3:79", "network": {"id": "aec9c0ce-a40e-4d0b-ad1d-2984fe3a4e89", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1834228009-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "349d3e2d0d8849819007c19528145c3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd0eb882-ab95-4373-aa20-ee565a9919e3", "external-id": "nsx-vlan-transportzone-510", "segmentation_id": 510, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5dabab4-c5", "ovs_interfaceid": "f5dabab4-c5af-42f4-bfdc-eec4330bdb0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 633.968295] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453194, 'name': Rename_Task, 'duration_secs': 0.206965} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.968506] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 633.968779] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f585816-5cc6-44f8-b547-41dea69683a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.979235] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 633.979235] env[65385]: value = "task-4453195" [ 633.979235] env[65385]: _type = "Task" [ 633.979235] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.989714] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453195, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.064015] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795d8bba-0eb5-4011-ae4a-69c6a84798c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.069995] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.074788] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e194da6d-6fb2-433d-ac7a-bda0463cbee5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.079910] env[65385]: DEBUG oslo_concurrency.lockutils [req-2355be1b-88c7-42aa-be7a-647f4f9b403c req-40f16adf-8cbe-4b53-bc27-2de9af98d568 service nova] Releasing lock "refresh_cache-a8c76329-5056-4cf0-ac32-85d46429d3db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.124534] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc2bb44-40b8-4515-a54a-88c298628b20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.138029] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9eb5b78-622a-421f-9a5a-46085eb587a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.157567] env[65385]: DEBUG nova.compute.provider_tree [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.227551] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "refresh_cache-1bc94ee7-d4f9-48c3-97f1-9e662eb50582" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.229391] env[65385]: DEBUG nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Instance network_info: |[{"id": "5f60e5e9-2860-4364-99c0-3c1433e4bddb", "address": "fa:16:3e:8e:64:4e", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f60e5e9-28", "ovs_interfaceid": "5f60e5e9-2860-4364-99c0-3c1433e4bddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 634.229491] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:64:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7514a465-f1a4-4a8b-b76b-726b1a9d7e2f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f60e5e9-2860-4364-99c0-3c1433e4bddb', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 634.241809] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 634.242258] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 634.243484] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7fb2ac0f-e0c4-49ab-a4bd-837d9da6054c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.272478] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 634.272478] env[65385]: value = "task-4453196" [ 634.272478] env[65385]: _type = "Task" [ 634.272478] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.284141] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453196, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.290528] env[65385]: DEBUG nova.network.neutron [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Successfully created port: 4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 634.319349] env[65385]: DEBUG nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 634.407107] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52194848-71af-ae06-2a7c-925b3a268f7c, 'name': SearchDatastore_Task, 'duration_secs': 0.012918} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.407901] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-838edb8c-a722-496b-bb81-c2b5a5e171aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.414363] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Releasing lock "refresh_cache-8ac44124-d8cf-421c-9ae4-c943df94550a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.414490] env[65385]: DEBUG nova.compute.manager [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Received event network-vif-plugged-82373c5b-5d69-406c-9fca-ed283c2a5f88 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 634.414688] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Acquiring lock "60067529-1071-4295-b1c9-21523bf347d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.414892] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Lock "60067529-1071-4295-b1c9-21523bf347d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 634.415828] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Lock "60067529-1071-4295-b1c9-21523bf347d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 634.415828] env[65385]: DEBUG nova.compute.manager [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] No waiting events found dispatching network-vif-plugged-82373c5b-5d69-406c-9fca-ed283c2a5f88 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 634.415828] env[65385]: WARNING nova.compute.manager [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Received unexpected event network-vif-plugged-82373c5b-5d69-406c-9fca-ed283c2a5f88 for instance with vm_state building and task_state spawning. [ 634.415828] env[65385]: DEBUG nova.compute.manager [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Received event network-changed-82373c5b-5d69-406c-9fca-ed283c2a5f88 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 634.415828] env[65385]: DEBUG nova.compute.manager [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Refreshing instance network info cache due to event network-changed-82373c5b-5d69-406c-9fca-ed283c2a5f88. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 634.416086] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Acquiring lock "refresh_cache-60067529-1071-4295-b1c9-21523bf347d9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.416086] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Acquired lock "refresh_cache-60067529-1071-4295-b1c9-21523bf347d9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.416282] env[65385]: DEBUG nova.network.neutron [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Refreshing network info cache for port 82373c5b-5d69-406c-9fca-ed283c2a5f88 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 634.417906] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 634.417906] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52874913-40d1-69fe-feca-bb3c356ea33b" [ 634.417906] env[65385]: _type = "Task" [ 634.417906] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.427338] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52874913-40d1-69fe-feca-bb3c356ea33b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.492983] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453195, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.662520] env[65385]: DEBUG nova.scheduler.client.report [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 634.704169] env[65385]: DEBUG nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 634.737882] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:47:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='195238370',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-503137054',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 634.738728] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 634.738728] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 634.738923] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 634.738923] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 634.739081] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 634.739318] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.739945] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 634.739945] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 634.739945] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 634.739945] env[65385]: DEBUG nova.virt.hardware [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 634.740930] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3eaa39-923d-45af-abfc-4988ce4384af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.750026] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ef54fb-28d0-49b5-965b-8d743fa921d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.784417] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453196, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.849759] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.919841] env[65385]: WARNING neutronclient.v2_0.client [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.920554] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.920899] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.939182] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52874913-40d1-69fe-feca-bb3c356ea33b, 'name': SearchDatastore_Task, 'duration_secs': 0.032749} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.939476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.939791] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 59450963-d83f-46e8-8c13-05d4f1818c64/59450963-d83f-46e8-8c13-05d4f1818c64.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 634.940139] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc6567d1-f5e0-4664-a1a7-9a73ea2e1f54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.948108] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 634.948108] env[65385]: value = "task-4453197" [ 634.948108] env[65385]: _type = "Task" [ 634.948108] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.957668] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453197, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.990331] env[65385]: DEBUG oslo_vmware.api [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453195, 'name': PowerOnVM_Task, 'duration_secs': 0.879492} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.990624] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.990830] env[65385]: INFO nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Took 12.00 seconds to spawn the instance on the hypervisor. [ 634.990997] env[65385]: DEBUG nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 634.991823] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb3656a-1a1a-420e-b58a-469c1c4f600e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.170523] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.171043] env[65385]: DEBUG nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 635.181044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.190s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.181267] env[65385]: DEBUG nova.objects.instance [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 635.289409] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453196, 'name': CreateVM_Task, 'duration_secs': 0.581179} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.289857] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 635.290159] env[65385]: WARNING neutronclient.v2_0.client [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.290584] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.290728] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.291051] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 635.292056] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f2faa1-2bd2-459c-95d0-d1af03e60c7b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.297907] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 635.297907] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dcd749-228b-6e1b-f339-383c0b51b906" [ 635.297907] env[65385]: _type = "Task" [ 635.297907] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.310401] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dcd749-228b-6e1b-f339-383c0b51b906, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.463882] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453197, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.510275] env[65385]: INFO nova.compute.manager [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Took 22.32 seconds to build instance. [ 635.690210] env[65385]: DEBUG nova.compute.utils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 635.692070] env[65385]: DEBUG nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 635.692314] env[65385]: DEBUG nova.network.neutron [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 635.692675] env[65385]: WARNING neutronclient.v2_0.client [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.693020] env[65385]: WARNING neutronclient.v2_0.client [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.693704] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.694250] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.809151] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dcd749-228b-6e1b-f339-383c0b51b906, 'name': SearchDatastore_Task, 'duration_secs': 0.034632} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.809634] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.809919] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 635.810221] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.810426] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.810705] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 635.811071] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c94f5a2f-203a-4305-9e0a-f0016a52e9b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.821209] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 635.821508] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 635.822404] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b22f01b3-0e3a-4fbb-bb0f-9f7d4aaad236 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.829367] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 635.829367] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52572e0d-4ce5-173a-fc1b-48e31d14cca5" [ 635.829367] env[65385]: _type = "Task" [ 635.829367] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.840364] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52572e0d-4ce5-173a-fc1b-48e31d14cca5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.965208] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453197, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.783135} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.965208] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 59450963-d83f-46e8-8c13-05d4f1818c64/59450963-d83f-46e8-8c13-05d4f1818c64.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.965208] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 635.965208] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1beb0d4a-389d-4512-b527-464339932570 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.975661] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 635.975661] env[65385]: value = "task-4453198" [ 635.975661] env[65385]: _type = "Task" [ 635.975661] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.987227] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453198, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.012476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bf992a7f-2128-4154-b7bc-91acf9fde42d tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "60067529-1071-4295-b1c9-21523bf347d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.599s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.082671] env[65385]: DEBUG nova.policy [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ba4783b423a4120a1035fabb02cd0aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9a1bf7c2dad4808a3782a3fba8a25fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 636.193990] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b74c011-e90f-4085-a0f4-999f10c17e76 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.196392] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 16.454s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.203621] env[65385]: DEBUG nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 636.242739] env[65385]: DEBUG nova.network.neutron [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Successfully updated port: 4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 636.342622] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52572e0d-4ce5-173a-fc1b-48e31d14cca5, 'name': SearchDatastore_Task, 'duration_secs': 0.017843} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.343895] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34841f69-8b15-4ac4-a783-1a8b38efa544 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.352038] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 636.352038] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524ed4d3-5f77-d470-650f-48a9e396ccf2" [ 636.352038] env[65385]: _type = "Task" [ 636.352038] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.369019] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524ed4d3-5f77-d470-650f-48a9e396ccf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.486238] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453198, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.224097} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.486864] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 636.487889] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a6c787-258c-4a66-959a-bb397ac94358 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.515361] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] 59450963-d83f-46e8-8c13-05d4f1818c64/59450963-d83f-46e8-8c13-05d4f1818c64.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 636.516185] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df4226a2-6e21-4c1c-ac95-1a843ef7ed7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.545163] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 636.545163] env[65385]: value = "task-4453199" [ 636.545163] env[65385]: _type = "Task" [ 636.545163] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.548031] env[65385]: DEBUG nova.compute.manager [None req-ad025213-6049-42f1-9859-3d1a6c19c18b tempest-ServerDiagnosticsV248Test-718726699 tempest-ServerDiagnosticsV248Test-718726699-project-admin] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 636.553821] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ddd719-a35c-4618-853e-f77c0dd94fab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.569457] env[65385]: INFO nova.compute.manager [None req-ad025213-6049-42f1-9859-3d1a6c19c18b tempest-ServerDiagnosticsV248Test-718726699 tempest-ServerDiagnosticsV248Test-718726699-project-admin] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Retrieving diagnostics [ 636.569783] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453199, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.570942] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f13fd95-900d-4edc-8343-d7fcc43620b3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.708162] env[65385]: INFO nova.compute.claims [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.748921] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.748921] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.748921] env[65385]: DEBUG nova.network.neutron [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 636.801255] env[65385]: DEBUG nova.network.neutron [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Successfully created port: 2206d06f-505b-4211-9823-29d900820228 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 636.819060] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.819475] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.865410] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524ed4d3-5f77-d470-650f-48a9e396ccf2, 'name': SearchDatastore_Task, 'duration_secs': 0.013843} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.865619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.866992] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 1bc94ee7-d4f9-48c3-97f1-9e662eb50582/1bc94ee7-d4f9-48c3-97f1-9e662eb50582.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.866992] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a05476d-d9b4-4c8b-ba66-bc8be01716f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.874375] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 636.874375] env[65385]: value = "task-4453200" [ 636.874375] env[65385]: _type = "Task" [ 636.874375] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.884166] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453200, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.065780] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453199, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.141115] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.141115] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.220547] env[65385]: INFO nova.compute.resource_tracker [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating resource usage from migration cbca16ba-15ec-46bc-8a77-271b74b9f582 [ 637.232239] env[65385]: DEBUG nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 637.255906] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.255906] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.276925] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 637.278021] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 637.278021] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 637.278021] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 637.278021] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 637.278021] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 637.278345] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.278345] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 637.278514] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 637.278666] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 637.278856] env[65385]: DEBUG nova.virt.hardware [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 637.279847] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3b7d70-6f1d-4a72-a1b6-751f7a89b305 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.293211] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b3bca9-f745-4f53-b133-eedf2595579b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.388340] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453200, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.399152] env[65385]: WARNING neutronclient.v2_0.client [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.401058] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.401058] env[65385]: WARNING openstack [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.433366] env[65385]: DEBUG nova.network.neutron [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 637.558651] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453199, 'name': ReconfigVM_Task, 'duration_secs': 0.772933} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.562402] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Reconfigured VM instance instance-00000011 to attach disk [datastore2] 59450963-d83f-46e8-8c13-05d4f1818c64/59450963-d83f-46e8-8c13-05d4f1818c64.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 637.563429] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c65e52f0-77b0-4c6e-8f80-956aca97c13e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.572591] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 637.572591] env[65385]: value = "task-4453201" [ 637.572591] env[65385]: _type = "Task" [ 637.572591] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.588714] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453201, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.645492] env[65385]: DEBUG nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 637.654491] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf59a64-652b-4b5b-8cda-f2a77767479f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.663350] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec58be6-eca9-4737-8b6e-7b111073cf21 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.702326] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "f77d861a-ae66-44c3-901d-e499926d6f5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.702667] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.702888] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "f77d861a-ae66-44c3-901d-e499926d6f5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.703100] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.703267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 637.705996] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac14a76-c8dd-42c5-b600-b171dc7a7192 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.709344] env[65385]: INFO nova.compute.manager [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Terminating instance [ 637.718173] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bf30b9-1ba4-4dbf-a482-76cdc772e61d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.738780] env[65385]: DEBUG nova.compute.provider_tree [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.887400] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453200, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693922} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.887699] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 1bc94ee7-d4f9-48c3-97f1-9e662eb50582/1bc94ee7-d4f9-48c3-97f1-9e662eb50582.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 637.887915] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 637.888211] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f14fd95-d7d1-46a2-a649-8b2d0d3decf2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.897135] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 637.897135] env[65385]: value = "task-4453202" [ 637.897135] env[65385]: _type = "Task" [ 637.897135] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.906809] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453202, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.085626] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453201, 'name': Rename_Task, 'duration_secs': 0.206153} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.085626] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 638.085860] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-368b8349-3f21-4a95-a105-c50438c11250 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.093602] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 638.093602] env[65385]: value = "task-4453203" [ 638.093602] env[65385]: _type = "Task" [ 638.093602] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.103117] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453203, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.180913] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.221675] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.222129] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.235187] env[65385]: DEBUG nova.compute.manager [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 638.235630] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.239183] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351cf94b-a246-4876-8fe1-e342b5165168 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.242802] env[65385]: DEBUG nova.scheduler.client.report [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 638.253519] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 638.253720] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63894290-a2c2-44f9-a6f6-dd60243b2a1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.262966] env[65385]: DEBUG oslo_vmware.api [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 638.262966] env[65385]: value = "task-4453204" [ 638.262966] env[65385]: _type = "Task" [ 638.262966] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.276062] env[65385]: DEBUG oslo_vmware.api [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.414663] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453202, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078735} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.415280] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 638.418609] env[65385]: DEBUG nova.network.neutron [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Updated VIF entry in instance network info cache for port 82373c5b-5d69-406c-9fca-ed283c2a5f88. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 638.418609] env[65385]: DEBUG nova.network.neutron [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Updating instance_info_cache with network_info: [{"id": "82373c5b-5d69-406c-9fca-ed283c2a5f88", "address": "fa:16:3e:4c:47:90", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82373c5b-5d", "ovs_interfaceid": "82373c5b-5d69-406c-9fca-ed283c2a5f88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.420137] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd03e33e-6208-4939-964d-983f51de0c01 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.449480] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 1bc94ee7-d4f9-48c3-97f1-9e662eb50582/1bc94ee7-d4f9-48c3-97f1-9e662eb50582.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 638.454982] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b5f5a4f-e081-4da1-ad9c-6261c4ce27d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.480786] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 638.480786] env[65385]: value = "task-4453205" [ 638.480786] env[65385]: _type = "Task" [ 638.480786] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.491183] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453205, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.528909] env[65385]: DEBUG nova.network.neutron [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Successfully updated port: 2206d06f-505b-4211-9823-29d900820228 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 638.536139] env[65385]: WARNING neutronclient.v2_0.client [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.536622] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.536996] env[65385]: WARNING openstack [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.614867] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453203, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.648988] env[65385]: DEBUG nova.network.neutron [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Updating instance_info_cache with network_info: [{"id": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "address": "fa:16:3e:89:46:11", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4343bbc8-b6", "ovs_interfaceid": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.748501] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.552s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.748724] env[65385]: INFO nova.compute.manager [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Migrating [ 638.748946] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.749088] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "compute-rpcapi-router" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.750909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.230s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.755421] env[65385]: INFO nova.compute.claims [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.761198] env[65385]: INFO nova.compute.rpcapi [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Automatically selected compute RPC version 6.4 from minimum service version 70 [ 638.765141] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "compute-rpcapi-router" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.799939] env[65385]: DEBUG oslo_vmware.api [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453204, 'name': PowerOffVM_Task, 'duration_secs': 0.418878} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.800387] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 638.800597] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 638.800977] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f85765ed-324f-496b-93ac-eac0493306bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.888485] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 638.888722] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 638.888894] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleting the datastore file [datastore2] f77d861a-ae66-44c3-901d-e499926d6f5b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 638.889202] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec6bbae8-3161-4eb1-bba7-6fe9882aaa4f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.897355] env[65385]: DEBUG oslo_vmware.api [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 638.897355] env[65385]: value = "task-4453207" [ 638.897355] env[65385]: _type = "Task" [ 638.897355] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.907772] env[65385]: DEBUG oslo_vmware.api [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453207, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.924390] env[65385]: DEBUG oslo_concurrency.lockutils [req-2c863fa8-76db-4bfd-bdd7-ccc837641d2e req-3b55e439-3e1a-4271-93b6-487c090b4199 service nova] Releasing lock "refresh_cache-60067529-1071-4295-b1c9-21523bf347d9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.995574] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453205, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.047175] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "refresh_cache-eb9d0510-b453-4695-9e1d-731217b9f8ec" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.047175] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "refresh_cache-eb9d0510-b453-4695-9e1d-731217b9f8ec" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.047416] env[65385]: DEBUG nova.network.neutron [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 639.113939] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453203, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.152546] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Releasing lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.152930] env[65385]: DEBUG nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Instance network_info: |[{"id": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "address": "fa:16:3e:89:46:11", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4343bbc8-b6", "ovs_interfaceid": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 639.153871] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:46:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f68ebd2a-3c68-48db-8c32-8a01497fc2e7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4343bbc8-b6b2-4198-9f4c-d6da5ff242dc', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.165585] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Creating folder: Project (2de8098800694299aae5aa2f59f405bb). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.165585] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5cf57303-860a-4e1b-8114-dccff094290f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.171797] env[65385]: DEBUG nova.compute.manager [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Received event network-vif-plugged-5f60e5e9-2860-4364-99c0-3c1433e4bddb {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 639.172152] env[65385]: DEBUG oslo_concurrency.lockutils [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Acquiring lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.172758] env[65385]: DEBUG oslo_concurrency.lockutils [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.173117] env[65385]: DEBUG oslo_concurrency.lockutils [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.173393] env[65385]: DEBUG nova.compute.manager [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] No waiting events found dispatching network-vif-plugged-5f60e5e9-2860-4364-99c0-3c1433e4bddb {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 639.173746] env[65385]: WARNING nova.compute.manager [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Received unexpected event network-vif-plugged-5f60e5e9-2860-4364-99c0-3c1433e4bddb for instance with vm_state building and task_state spawning. [ 639.174242] env[65385]: DEBUG nova.compute.manager [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Received event network-changed-5f60e5e9-2860-4364-99c0-3c1433e4bddb {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 639.174434] env[65385]: DEBUG nova.compute.manager [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Refreshing instance network info cache due to event network-changed-5f60e5e9-2860-4364-99c0-3c1433e4bddb. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 639.174720] env[65385]: DEBUG oslo_concurrency.lockutils [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Acquiring lock "refresh_cache-1bc94ee7-d4f9-48c3-97f1-9e662eb50582" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.175261] env[65385]: DEBUG oslo_concurrency.lockutils [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Acquired lock "refresh_cache-1bc94ee7-d4f9-48c3-97f1-9e662eb50582" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.175261] env[65385]: DEBUG nova.network.neutron [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Refreshing network info cache for port 5f60e5e9-2860-4364-99c0-3c1433e4bddb {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 639.181715] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Created folder: Project (2de8098800694299aae5aa2f59f405bb) in parent group-v870881. [ 639.182094] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Creating folder: Instances. Parent ref: group-v870934. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.183380] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7c5eb21-656c-4d34-ba9d-f17f0aa878d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.198826] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Created folder: Instances in parent group-v870934. [ 639.200687] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 639.202375] env[65385]: DEBUG nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Received event network-changed-7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 639.202666] env[65385]: DEBUG nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Refreshing instance network info cache due to event network-changed-7d3f8bf5-22d2-41b0-9743-0616daaef615. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 639.202968] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Acquiring lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.203196] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Acquired lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.204454] env[65385]: DEBUG nova.network.neutron [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Refreshing network info cache for port 7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 639.205946] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 639.205946] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8af73f7-659f-4a85-9452-06686e4cc24e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.230780] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.230780] env[65385]: value = "task-4453210" [ 639.230780] env[65385]: _type = "Task" [ 639.230780] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.241865] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453210, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.289035] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.289360] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.289707] env[65385]: DEBUG nova.network.neutron [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 639.409613] env[65385]: DEBUG oslo_vmware.api [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453207, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289918} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.409816] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 639.410056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 639.410196] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.410332] env[65385]: INFO nova.compute.manager [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Took 1.18 seconds to destroy the instance on the hypervisor. [ 639.410611] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 639.410793] env[65385]: DEBUG nova.compute.manager [-] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 639.410836] env[65385]: DEBUG nova.network.neutron [-] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 639.411423] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.412404] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.413250] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.492932] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453205, 'name': ReconfigVM_Task, 'duration_secs': 0.6748} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.493227] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 1bc94ee7-d4f9-48c3-97f1-9e662eb50582/1bc94ee7-d4f9-48c3-97f1-9e662eb50582.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.494089] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ecc5644-7049-4507-afcd-50a3bdec91bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.502489] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 639.502489] env[65385]: value = "task-4453211" [ 639.502489] env[65385]: _type = "Task" [ 639.502489] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.509755] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.518106] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453211, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.554524] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.555192] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.613083] env[65385]: DEBUG oslo_vmware.api [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453203, 'name': PowerOnVM_Task, 'duration_secs': 1.025927} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.613418] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 639.613755] env[65385]: INFO nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Took 12.02 seconds to spawn the instance on the hypervisor. [ 639.614157] env[65385]: DEBUG nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 639.615447] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322cdce6-6c86-46ea-9751-ee7fb5e88190 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.683045] env[65385]: WARNING neutronclient.v2_0.client [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.683937] env[65385]: WARNING openstack [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.684194] env[65385]: WARNING openstack [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.706730] env[65385]: WARNING neutronclient.v2_0.client [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.707339] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.708264] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.742528] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453210, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.761532] env[65385]: DEBUG nova.network.neutron [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 639.794053] env[65385]: WARNING neutronclient.v2_0.client [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.794053] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.794053] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.917971] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.917971] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.013347] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453211, 'name': Rename_Task, 'duration_secs': 0.258438} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.016541] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 640.017049] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f4a61e6-b697-4276-b19d-cfd4fe5fd608 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.025722] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 640.025722] env[65385]: value = "task-4453212" [ 640.025722] env[65385]: _type = "Task" [ 640.025722] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.043956] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453212, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.087255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.087442] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.142485] env[65385]: INFO nova.compute.manager [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Took 25.84 seconds to build instance. [ 640.192541] env[65385]: WARNING openstack [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.192541] env[65385]: WARNING openstack [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.243525] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453210, 'name': CreateVM_Task, 'duration_secs': 0.537617} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.243843] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 640.245669] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.246124] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.253713] env[65385]: WARNING neutronclient.v2_0.client [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.255597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.255597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.255597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 640.257909] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-677033f6-a19e-4184-81dc-205734b48f1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.260803] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04ddda3-56ce-4fdb-b446-ad1af10bfbd5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.271865] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db0a67b-cf85-4c02-b9d6-427ce93ffcf0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.275727] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 640.275727] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521b2102-ae5e-2c98-5710-cf5b1f27c3fe" [ 640.275727] env[65385]: _type = "Task" [ 640.275727] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.284266] env[65385]: WARNING neutronclient.v2_0.client [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.284991] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.285402] env[65385]: WARNING openstack [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.338416] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290db4a1-2fe6-4823-95ff-570a376263d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.348353] env[65385]: DEBUG nova.network.neutron [-] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 640.349426] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521b2102-ae5e-2c98-5710-cf5b1f27c3fe, 'name': SearchDatastore_Task, 'duration_secs': 0.012275} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.351964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.351964] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.351964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.351964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.352223] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.352361] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e6c7967-3039-4a16-99d3-9c963cc70f67 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.359237] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4389af24-15dc-4391-8961-3f74674eb519 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.368091] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.368302] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 640.377870] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da0dc59f-1f7d-4d2d-9fca-f6b744b864d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.380995] env[65385]: DEBUG nova.compute.provider_tree [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.388064] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 640.388064] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5295c044-e546-b599-0cdb-dff718c91899" [ 640.388064] env[65385]: _type = "Task" [ 640.388064] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.401291] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "c5b85574-df14-4a7a-ada0-c9668bf18a33" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.401709] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "c5b85574-df14-4a7a-ada0-c9668bf18a33" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.402047] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "c5b85574-df14-4a7a-ada0-c9668bf18a33-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.402387] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "c5b85574-df14-4a7a-ada0-c9668bf18a33-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.402666] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "c5b85574-df14-4a7a-ada0-c9668bf18a33-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.405433] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5295c044-e546-b599-0cdb-dff718c91899, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.406067] env[65385]: INFO nova.compute.manager [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Terminating instance [ 640.523537] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.523537] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.542587] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453212, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.589728] env[65385]: DEBUG nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 640.645733] env[65385]: DEBUG oslo_concurrency.lockutils [None req-835c6a67-b00f-4240-90dc-a0f429fd210c tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "59450963-d83f-46e8-8c13-05d4f1818c64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.829s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.822499] env[65385]: WARNING neutronclient.v2_0.client [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.823444] env[65385]: WARNING openstack [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.823949] env[65385]: WARNING openstack [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.843869] env[65385]: WARNING neutronclient.v2_0.client [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.844662] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.845110] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.854344] env[65385]: INFO nova.compute.manager [-] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Took 1.44 seconds to deallocate network for instance. [ 640.856896] env[65385]: DEBUG nova.network.neutron [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Updating instance_info_cache with network_info: [{"id": "2206d06f-505b-4211-9823-29d900820228", "address": "fa:16:3e:81:0f:eb", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2206d06f-50", "ovs_interfaceid": "2206d06f-505b-4211-9823-29d900820228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 640.876860] env[65385]: WARNING neutronclient.v2_0.client [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.877605] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.878045] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.888474] env[65385]: DEBUG nova.scheduler.client.report [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 640.908333] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5295c044-e546-b599-0cdb-dff718c91899, 'name': SearchDatastore_Task, 'duration_secs': 0.010994} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.908333] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d230d1df-a4ad-4fd9-bcee-854bc6ae8da9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.910276] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "refresh_cache-c5b85574-df14-4a7a-ada0-c9668bf18a33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.910435] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquired lock "refresh_cache-c5b85574-df14-4a7a-ada0-c9668bf18a33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.910608] env[65385]: DEBUG nova.network.neutron [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 640.916411] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 640.916411] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527fd885-6cfe-269f-0943-bd755d173a2c" [ 640.916411] env[65385]: _type = "Task" [ 640.916411] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.926829] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527fd885-6cfe-269f-0943-bd755d173a2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.043661] env[65385]: DEBUG oslo_vmware.api [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453212, 'name': PowerOnVM_Task, 'duration_secs': 0.727804} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.044093] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 641.044528] env[65385]: INFO nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Took 10.89 seconds to spawn the instance on the hypervisor. [ 641.044963] env[65385]: DEBUG nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 641.045851] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee212166-5bc5-4c08-89ca-42bd7c4c4c47 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.121477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.278631] env[65385]: DEBUG nova.network.neutron [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Updated VIF entry in instance network info cache for port 5f60e5e9-2860-4364-99c0-3c1433e4bddb. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 641.279103] env[65385]: DEBUG nova.network.neutron [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Updating instance_info_cache with network_info: [{"id": "5f60e5e9-2860-4364-99c0-3c1433e4bddb", "address": "fa:16:3e:8e:64:4e", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f60e5e9-28", "ovs_interfaceid": "5f60e5e9-2860-4364-99c0-3c1433e4bddb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 641.361031] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "refresh_cache-eb9d0510-b453-4695-9e1d-731217b9f8ec" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.361436] env[65385]: DEBUG nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance network_info: |[{"id": "2206d06f-505b-4211-9823-29d900820228", "address": "fa:16:3e:81:0f:eb", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2206d06f-50", "ovs_interfaceid": "2206d06f-505b-4211-9823-29d900820228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 641.362507] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:0f:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6815237d-f565-474d-a3c0-9c675478eb00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2206d06f-505b-4211-9823-29d900820228', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 641.373822] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating folder: Project (a9a1bf7c2dad4808a3782a3fba8a25fc). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 641.375738] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.376461] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b2ae6e3-de27-4a08-8037-279c4fde1440 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.389665] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created folder: Project (a9a1bf7c2dad4808a3782a3fba8a25fc) in parent group-v870881. [ 641.389962] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating folder: Instances. Parent ref: group-v870937. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 641.390330] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19c35b01-552a-412e-befb-2b169ccd3b14 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.395654] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.644s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.395654] env[65385]: DEBUG nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 641.398231] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.799s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 641.398732] env[65385]: DEBUG nova.objects.instance [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lazy-loading 'resources' on Instance uuid a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 641.404754] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created folder: Instances in parent group-v870937. [ 641.405627] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 641.405627] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 641.405627] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2172dd3d-78ab-4283-aae9-6c93b3770e5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.430920] env[65385]: DEBUG nova.network.neutron [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updated VIF entry in instance network info cache for port 7d3f8bf5-22d2-41b0-9743-0616daaef615. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 641.431536] env[65385]: DEBUG nova.network.neutron [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updating instance_info_cache with network_info: [{"id": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "address": "fa:16:3e:f9:84:04", "network": {"id": "9ed70dba-8022-422b-8af1-9e125786bbe0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-823001245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73c43e9322314241811db1a37e18f9b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f8bf5-22", "ovs_interfaceid": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 641.433139] env[65385]: WARNING neutronclient.v2_0.client [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.434087] env[65385]: WARNING openstack [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.435278] env[65385]: WARNING openstack [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.455053] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527fd885-6cfe-269f-0943-bd755d173a2c, 'name': SearchDatastore_Task, 'duration_secs': 0.028577} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.457138] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.457450] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 427d5c6d-ab01-42ef-8ee5-edd608896b8b/427d5c6d-ab01-42ef-8ee5-edd608896b8b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 641.459040] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 641.459040] env[65385]: value = "task-4453215" [ 641.459040] env[65385]: _type = "Task" [ 641.459040] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.460405] env[65385]: DEBUG nova.network.neutron [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance_info_cache with network_info: [{"id": "03771ba6-a616-4715-afb9-19306b1b0903", "address": "fa:16:3e:2b:4a:04", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03771ba6-a6", "ovs_interfaceid": "03771ba6-a616-4715-afb9-19306b1b0903", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 641.461693] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce7ea5ba-d825-4406-97aa-654d5b7fe307 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.480457] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453215, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.482361] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 641.482361] env[65385]: value = "task-4453216" [ 641.482361] env[65385]: _type = "Task" [ 641.482361] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.494648] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453216, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.495987] env[65385]: DEBUG nova.network.neutron [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 641.566422] env[65385]: INFO nova.compute.manager [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Took 27.09 seconds to build instance. [ 641.575221] env[65385]: DEBUG nova.network.neutron [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 641.785498] env[65385]: DEBUG oslo_concurrency.lockutils [req-2243ae2a-e0a8-4ff9-82b7-99b03776dfe5 req-33d4326e-95ce-463b-afec-2c326f27ddf8 service nova] Releasing lock "refresh_cache-1bc94ee7-d4f9-48c3-97f1-9e662eb50582" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.902193] env[65385]: DEBUG nova.compute.utils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 641.907506] env[65385]: DEBUG nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 641.907796] env[65385]: DEBUG nova.network.neutron [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 641.908707] env[65385]: WARNING neutronclient.v2_0.client [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.908948] env[65385]: WARNING neutronclient.v2_0.client [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.910068] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.910640] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.944506] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Releasing lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.945388] env[65385]: DEBUG nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Received event network-vif-deleted-a9fd2b9d-1995-43d9-8550-a27e7828eeb6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 641.945885] env[65385]: DEBUG nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Received event network-vif-plugged-97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 641.946201] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Acquiring lock "59450963-d83f-46e8-8c13-05d4f1818c64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.946790] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Lock "59450963-d83f-46e8-8c13-05d4f1818c64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 641.946846] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Lock "59450963-d83f-46e8-8c13-05d4f1818c64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.947085] env[65385]: DEBUG nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] No waiting events found dispatching network-vif-plugged-97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 641.947257] env[65385]: WARNING nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Received unexpected event network-vif-plugged-97c1d2c0-12b5-428b-85b4-81b8305d87f9 for instance with vm_state building and task_state spawning. [ 641.947678] env[65385]: DEBUG nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Received event network-changed-97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 641.947678] env[65385]: DEBUG nova.compute.manager [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Refreshing instance network info cache due to event network-changed-97c1d2c0-12b5-428b-85b4-81b8305d87f9. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 641.947797] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Acquiring lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.947882] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Acquired lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.948073] env[65385]: DEBUG nova.network.neutron [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Refreshing network info cache for port 97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 641.972946] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.988905] env[65385]: DEBUG nova.policy [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ba4783b423a4120a1035fabb02cd0aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9a1bf7c2dad4808a3782a3fba8a25fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 642.004392] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453215, 'name': CreateVM_Task, 'duration_secs': 0.4428} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.005523] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 642.006599] env[65385]: WARNING neutronclient.v2_0.client [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.007152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.007242] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.007525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 642.010383] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e258d581-959e-44d6-aa67-32a65e1cf6ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.018032] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453216, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.023440] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 642.023440] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527a8c98-dd6b-75cf-e14b-1453bf7dcd81" [ 642.023440] env[65385]: _type = "Task" [ 642.023440] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.037229] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527a8c98-dd6b-75cf-e14b-1453bf7dcd81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.070191] env[65385]: DEBUG oslo_concurrency.lockutils [None req-025ddcc4-ae4a-4cfb-b8d3-b1a412a6d139 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.561s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.079582] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Releasing lock "refresh_cache-c5b85574-df14-4a7a-ada0-c9668bf18a33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.079728] env[65385]: DEBUG nova.compute.manager [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 642.079980] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 642.081218] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5391d9d6-ff6e-42e9-badc-72d1ba9be0fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.090878] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 642.091216] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb59a0dc-8a99-4bef-ac4c-d61c35e0c9d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.099544] env[65385]: DEBUG oslo_vmware.api [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 642.099544] env[65385]: value = "task-4453217" [ 642.099544] env[65385]: _type = "Task" [ 642.099544] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.115541] env[65385]: DEBUG oslo_vmware.api [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.229811] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "065659ae-2163-4d6b-a905-fc99d12b6790" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.229974] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "065659ae-2163-4d6b-a905-fc99d12b6790" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.230833] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "065659ae-2163-4d6b-a905-fc99d12b6790-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.230833] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "065659ae-2163-4d6b-a905-fc99d12b6790-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.230833] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "065659ae-2163-4d6b-a905-fc99d12b6790-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.233042] env[65385]: INFO nova.compute.manager [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Terminating instance [ 642.408502] env[65385]: DEBUG nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 642.436330] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bff343-a78b-4552-bccc-f24621365e0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.444572] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c3117a-2618-4d3b-9821-0645df5bae05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.453051] env[65385]: WARNING neutronclient.v2_0.client [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.453717] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.454087] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.490528] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95db6a9-c0fa-48a4-9a3b-9e37f33a3269 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.502085] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae22d653-0e8c-4b1f-8470-ae2ae7d5f19c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.510798] env[65385]: DEBUG nova.network.neutron [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Successfully created port: d59b902f-9fb1-4d56-96f9-6301024412fa {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 642.513422] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453216, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63861} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.515332] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 427d5c6d-ab01-42ef-8ee5-edd608896b8b/427d5c6d-ab01-42ef-8ee5-edd608896b8b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 642.515652] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.529136] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8eab917e-cd0e-4f92-af69-1fceb2aa56f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.530559] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "a71d2fa6-006c-4f76-8ae8-467e149297da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.530800] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.532219] env[65385]: DEBUG nova.compute.provider_tree [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.546051] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527a8c98-dd6b-75cf-e14b-1453bf7dcd81, 'name': SearchDatastore_Task, 'duration_secs': 0.034868} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.547552] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.547784] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 642.548040] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.548220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.548371] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 642.549552] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 642.549552] env[65385]: value = "task-4453218" [ 642.549552] env[65385]: _type = "Task" [ 642.549552] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.549552] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9d24b49-b496-4c39-9675-956045f3140e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.561916] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453218, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.565326] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 642.565506] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 642.566586] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6598e792-fad2-4d97-8604-8d5d846ae9d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.576160] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 642.576160] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52326bd3-13cd-2adf-7c8a-314587881ec0" [ 642.576160] env[65385]: _type = "Task" [ 642.576160] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.585472] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52326bd3-13cd-2adf-7c8a-314587881ec0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.614500] env[65385]: DEBUG oslo_vmware.api [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453217, 'name': PowerOffVM_Task, 'duration_secs': 0.20531} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.615186] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 642.615453] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 642.615857] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-529ef985-60d2-41e6-9fdb-204109e10da4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.643804] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 642.643942] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 642.644176] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Deleting the datastore file [datastore1] c5b85574-df14-4a7a-ada0-c9668bf18a33 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 642.644478] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f9d3428-cb48-4706-822e-85f9577b539b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.653225] env[65385]: DEBUG oslo_vmware.api [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for the task: (returnval){ [ 642.653225] env[65385]: value = "task-4453220" [ 642.653225] env[65385]: _type = "Task" [ 642.653225] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.662646] env[65385]: DEBUG oslo_vmware.api [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453220, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.741589] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "refresh_cache-065659ae-2163-4d6b-a905-fc99d12b6790" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.741882] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquired lock "refresh_cache-065659ae-2163-4d6b-a905-fc99d12b6790" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.742151] env[65385]: DEBUG nova.network.neutron [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 643.034347] env[65385]: DEBUG nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 643.039026] env[65385]: DEBUG nova.scheduler.client.report [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 643.064984] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453218, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.168385} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.065588] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 643.068045] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebfba9f-b7dd-4880-a4e1-d43b48ddf3d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.095069] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Reconfiguring VM instance instance-00000013 to attach disk [datastore2] 427d5c6d-ab01-42ef-8ee5-edd608896b8b/427d5c6d-ab01-42ef-8ee5-edd608896b8b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.098887] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-615b96b0-bd33-417d-bf0f-914a552bcda3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.124351] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52326bd3-13cd-2adf-7c8a-314587881ec0, 'name': SearchDatastore_Task, 'duration_secs': 0.035749} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.126595] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 643.126595] env[65385]: value = "task-4453221" [ 643.126595] env[65385]: _type = "Task" [ 643.126595] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.126872] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85990441-19a5-4bd9-90cc-c4d02227a2da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.138102] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 643.138102] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d150b1-7a42-4fe9-62db-ec2abd144096" [ 643.138102] env[65385]: _type = "Task" [ 643.138102] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.141568] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.151500] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d150b1-7a42-4fe9-62db-ec2abd144096, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.162738] env[65385]: DEBUG oslo_vmware.api [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Task: {'id': task-4453220, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228928} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.163458] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 643.163458] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 643.163458] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.163927] env[65385]: INFO nova.compute.manager [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Took 1.08 seconds to destroy the instance on the hypervisor. [ 643.163927] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 643.164228] env[65385]: DEBUG nova.compute.manager [-] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 643.164228] env[65385]: DEBUG nova.network.neutron [-] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 643.164479] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.165116] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.165194] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.245049] env[65385]: WARNING neutronclient.v2_0.client [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.246348] env[65385]: WARNING openstack [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.246874] env[65385]: WARNING openstack [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.418737] env[65385]: DEBUG nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 643.447523] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 643.447836] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 643.448018] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 643.448268] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 643.448433] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 643.448641] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 643.448859] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.449017] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 643.449180] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 643.449333] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 643.449506] env[65385]: DEBUG nova.virt.hardware [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 643.450378] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c3f8ec-8586-492d-a7d5-f44de4141deb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.459447] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8fa9a2-ff66-4af8-a14c-44296c007230 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.498208] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63bf9dc3-45a5-4e81-9008-d6da006865b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.517605] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance '42b9edf6-1873-49c3-8074-8eef654ac371' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 643.544413] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.146s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 643.546684] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.510s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.546906] env[65385]: DEBUG nova.objects.instance [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lazy-loading 'resources' on Instance uuid ef235d8e-6193-4010-ac50-72cd275e965a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 643.561969] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.571129] env[65385]: INFO nova.scheduler.client.report [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Deleted allocations for instance a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3 [ 643.640589] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453221, 'name': ReconfigVM_Task, 'duration_secs': 0.50761} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.640589] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Reconfigured VM instance instance-00000013 to attach disk [datastore2] 427d5c6d-ab01-42ef-8ee5-edd608896b8b/427d5c6d-ab01-42ef-8ee5-edd608896b8b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 643.640957] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0dd7caa1-60dd-4f79-bc8e-c8068f38d6b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.652225] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d150b1-7a42-4fe9-62db-ec2abd144096, 'name': SearchDatastore_Task, 'duration_secs': 0.032858} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.653558] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.653817] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 643.655448] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 643.655448] env[65385]: value = "task-4453222" [ 643.655448] env[65385]: _type = "Task" [ 643.655448] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.655448] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-471d30c1-388e-4cee-ab60-22e3908da182 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.668447] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453222, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.670052] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 643.670052] env[65385]: value = "task-4453223" [ 643.670052] env[65385]: _type = "Task" [ 643.670052] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.680110] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453223, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.723172] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.723656] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.813611] env[65385]: DEBUG nova.network.neutron [-] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 643.813885] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.831455] env[65385]: DEBUG nova.network.neutron [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 644.024829] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 644.024829] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2bd2c400-4a52-4041-85c1-54396e69afba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.034941] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 644.034941] env[65385]: value = "task-4453224" [ 644.034941] env[65385]: _type = "Task" [ 644.034941] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.047490] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.085202] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed23ebf1-15bb-481a-83bb-55d57f90ff6c tempest-TenantUsagesTestJSON-1602940447 tempest-TenantUsagesTestJSON-1602940447-project-member] Lock "a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.450s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.171588] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453222, 'name': Rename_Task, 'duration_secs': 0.215261} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.175142] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.175521] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ecb1609-112f-4c92-8f71-0ac11ab3b0f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.189661] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453223, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.192079] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 644.192079] env[65385]: value = "task-4453225" [ 644.192079] env[65385]: _type = "Task" [ 644.192079] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.204655] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453225, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.260904] env[65385]: DEBUG nova.network.neutron [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Successfully updated port: d59b902f-9fb1-4d56-96f9-6301024412fa {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 644.316456] env[65385]: DEBUG nova.network.neutron [-] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 644.415718] env[65385]: WARNING neutronclient.v2_0.client [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.416746] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.417298] env[65385]: WARNING openstack [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.464844] env[65385]: DEBUG nova.network.neutron [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 644.535936] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1bdf77-b031-4b06-9b53-2fa8c16f079d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.549663] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35815d97-47a9-4529-8016-cc046c71471c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.557106] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453224, 'name': PowerOffVM_Task, 'duration_secs': 0.497296} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.557546] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 644.557726] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance '42b9edf6-1873-49c3-8074-8eef654ac371' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 644.589934] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5e936f-a545-4fab-8423-056ee8e8a560 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.595836] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.596077] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.599615] env[65385]: DEBUG nova.network.neutron [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Updated VIF entry in instance network info cache for port 97c1d2c0-12b5-428b-85b4-81b8305d87f9. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 644.599615] env[65385]: DEBUG nova.network.neutron [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Updating instance_info_cache with network_info: [{"id": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "address": "fa:16:3e:3a:2c:78", "network": {"id": "a2c91019-c13d-4581-aa0b-05553afc8f57", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1163204339-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f858126fa23d43cbafd1b677206f28ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97c1d2c0-12", "ovs_interfaceid": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 644.607026] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adaa69bf-da6d-4aeb-ae41-fb8e06911d48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.623641] env[65385]: DEBUG nova.compute.provider_tree [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.682636] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453223, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.713672} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.683231] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 644.683231] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 644.683477] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cee1b94-d5c5-4568-a1d6-180fed3bec3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.692355] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 644.692355] env[65385]: value = "task-4453226" [ 644.692355] env[65385]: _type = "Task" [ 644.692355] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.709271] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453225, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.709550] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453226, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.766334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "refresh_cache-75ba706f-2e72-4f84-b02f-db4381951e77" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.766549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "refresh_cache-75ba706f-2e72-4f84-b02f-db4381951e77" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.766728] env[65385]: DEBUG nova.network.neutron [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 644.818971] env[65385]: INFO nova.compute.manager [-] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Took 1.65 seconds to deallocate network for instance. [ 644.968519] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Releasing lock "refresh_cache-065659ae-2163-4d6b-a905-fc99d12b6790" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 644.969087] env[65385]: DEBUG nova.compute.manager [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 644.969184] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.970151] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085983b8-d825-4646-8b4a-0b6b4f4dd13d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.980975] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 644.981847] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8d4c6bf-3577-45dc-a104-2b3f2750575c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.991946] env[65385]: DEBUG oslo_vmware.api [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 644.991946] env[65385]: value = "task-4453227" [ 644.991946] env[65385]: _type = "Task" [ 644.991946] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.003198] env[65385]: DEBUG oslo_vmware.api [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453227, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.093304] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 645.093708] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 645.094509] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 645.094823] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 645.095195] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 645.095412] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 645.095706] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.095943] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 645.096189] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 645.096440] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 645.096663] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 645.102733] env[65385]: DEBUG nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 645.106715] env[65385]: DEBUG oslo_concurrency.lockutils [req-24141f44-c5b4-47f8-8604-674b9fd6443d req-4e5d4e47-3a05-46d9-92c5-531b63e61c31 service nova] Releasing lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.107216] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e2e34b7-72fa-4496-b220-4d16a85cb7ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.128200] env[65385]: DEBUG nova.scheduler.client.report [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 645.136486] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 645.136486] env[65385]: value = "task-4453228" [ 645.136486] env[65385]: _type = "Task" [ 645.136486] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.150461] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453228, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.214584] env[65385]: DEBUG oslo_vmware.api [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453225, 'name': PowerOnVM_Task, 'duration_secs': 0.927793} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.219276] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.219910] env[65385]: INFO nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Took 10.52 seconds to spawn the instance on the hypervisor. [ 645.220018] env[65385]: DEBUG nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 645.220474] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453226, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07126} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.222305] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87bcc535-dc03-4034-bb12-80dc97893bd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.225695] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 645.227289] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef7e514-793c-4d0b-abe0-1a948e6909b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.258151] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 645.259795] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb7e590a-47d7-4328-bdf9-ab871b7c8430 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.278749] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.279157] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.295931] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 645.295931] env[65385]: value = "task-4453229" [ 645.295931] env[65385]: _type = "Task" [ 645.295931] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.307818] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453229, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.327356] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.371108] env[65385]: DEBUG nova.network.neutron [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 645.467423] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.467423] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.504435] env[65385]: DEBUG oslo_vmware.api [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453227, 'name': PowerOffVM_Task, 'duration_secs': 0.274291} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.504435] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 645.504435] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 645.504585] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2672c400-b186-4773-92b3-a140256d2d15 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.533736] env[65385]: DEBUG nova.compute.manager [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Received event network-changed-e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 645.533965] env[65385]: DEBUG nova.compute.manager [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Refreshing instance network info cache due to event network-changed-e37bf46a-308d-44a5-b573-64c3b7aa5a73. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 645.534246] env[65385]: DEBUG oslo_concurrency.lockutils [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Acquiring lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.534430] env[65385]: DEBUG oslo_concurrency.lockutils [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Acquired lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.534594] env[65385]: DEBUG nova.network.neutron [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Refreshing network info cache for port e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 645.538430] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 645.538430] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 645.538430] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Deleting the datastore file [datastore2] 065659ae-2163-4d6b-a905-fc99d12b6790 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 645.538430] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59ee525b-d3d8-42a1-8d4e-d1d6264ca3ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.548133] env[65385]: DEBUG oslo_vmware.api [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for the task: (returnval){ [ 645.548133] env[65385]: value = "task-4453231" [ 645.548133] env[65385]: _type = "Task" [ 645.548133] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.561616] env[65385]: DEBUG oslo_vmware.api [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.565770] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Received event network-vif-plugged-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 645.565770] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Acquiring lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.565955] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.566116] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.566398] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] No waiting events found dispatching network-vif-plugged-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 645.566669] env[65385]: WARNING nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Received unexpected event network-vif-plugged-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc for instance with vm_state active and task_state None. [ 645.566977] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Received event network-changed-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 645.567194] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Refreshing instance network info cache due to event network-changed-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 645.567476] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Acquiring lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.567693] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Acquired lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.567923] env[65385]: DEBUG nova.network.neutron [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Refreshing network info cache for port 4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 645.571738] env[65385]: WARNING neutronclient.v2_0.client [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.574809] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.574809] env[65385]: WARNING openstack [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.637073] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.090s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.640373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.342s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.644660] env[65385]: INFO nova.compute.claims [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.671838] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453228, 'name': ReconfigVM_Task, 'duration_secs': 0.291221} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.672467] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance '42b9edf6-1873-49c3-8074-8eef654ac371' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 645.680540] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.733326] env[65385]: INFO nova.scheduler.client.report [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Deleted allocations for instance ef235d8e-6193-4010-ac50-72cd275e965a [ 645.774761] env[65385]: DEBUG nova.network.neutron [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Updating instance_info_cache with network_info: [{"id": "d59b902f-9fb1-4d56-96f9-6301024412fa", "address": "fa:16:3e:f2:b1:5a", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd59b902f-9f", "ovs_interfaceid": "d59b902f-9fb1-4d56-96f9-6301024412fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 645.811261] env[65385]: INFO nova.compute.manager [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Took 29.82 seconds to build instance. [ 645.829229] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453229, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.039575] env[65385]: WARNING neutronclient.v2_0.client [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.040231] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.040623] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.064564] env[65385]: DEBUG oslo_vmware.api [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Task: {'id': task-4453231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205659} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.065913] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 646.066980] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 646.067340] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 646.067621] env[65385]: INFO nova.compute.manager [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Took 1.10 seconds to destroy the instance on the hypervisor. [ 646.067984] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 646.068321] env[65385]: DEBUG nova.compute.manager [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 646.068472] env[65385]: DEBUG nova.network.neutron [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 646.068874] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.069717] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.070426] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.082441] env[65385]: WARNING neutronclient.v2_0.client [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.083228] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.083869] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.183982] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 646.184389] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 646.184611] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 646.184890] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 646.189316] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 646.189747] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 646.190073] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.190518] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 646.190518] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 646.190746] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 646.190981] env[65385]: DEBUG nova.virt.hardware [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 646.201671] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Reconfiguring VM instance instance-00000006 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 646.203276] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8df921b-b108-46de-9686-e9bb0a083a65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.231254] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.231346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.239762] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 646.239762] env[65385]: value = "task-4453232" [ 646.239762] env[65385]: _type = "Task" [ 646.239762] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.249255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd83c4c9-f846-48b7-85ce-59c0c55ce8f5 tempest-VolumesAdminNegativeTest-558667950 tempest-VolumesAdminNegativeTest-558667950-project-member] Lock "ef235d8e-6193-4010-ac50-72cd275e965a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.132s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.251414] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453232, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.280745] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "refresh_cache-75ba706f-2e72-4f84-b02f-db4381951e77" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.281472] env[65385]: DEBUG nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Instance network_info: |[{"id": "d59b902f-9fb1-4d56-96f9-6301024412fa", "address": "fa:16:3e:f2:b1:5a", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd59b902f-9f", "ovs_interfaceid": "d59b902f-9fb1-4d56-96f9-6301024412fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 646.282169] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:b1:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6815237d-f565-474d-a3c0-9c675478eb00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd59b902f-9fb1-4d56-96f9-6301024412fa', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 646.294051] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 646.295217] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 646.295838] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b1e0449-4b20-44e5-ae43-c1fcf9b48e38 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.325231] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9c3f7515-e894-4795-b87b-fc6e20f5ebbb tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.235s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.341305] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453229, 'name': ReconfigVM_Task, 'duration_secs': 0.636796} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.343195] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Reconfigured VM instance instance-00000014 to attach disk [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 646.343973] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 646.343973] env[65385]: value = "task-4453233" [ 646.343973] env[65385]: _type = "Task" [ 646.343973] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.344274] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-692dd8e5-1da3-457e-862c-8293e1a3e4da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.357701] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453233, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.359488] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 646.359488] env[65385]: value = "task-4453234" [ 646.359488] env[65385]: _type = "Task" [ 646.359488] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.370829] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453234, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.734589] env[65385]: DEBUG nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 646.763841] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453232, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.800526] env[65385]: DEBUG nova.network.neutron [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 646.800526] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.861548] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453233, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.878588] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453234, 'name': Rename_Task, 'duration_secs': 0.247052} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.878907] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 646.879134] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-592babf4-83c1-4bf4-859e-3faf7de1df82 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.886781] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 646.886781] env[65385]: value = "task-4453235" [ 646.886781] env[65385]: _type = "Task" [ 646.886781] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.693088] env[65385]: DEBUG nova.network.neutron [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 647.698028] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453235, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.698028] env[65385]: WARNING oslo_vmware.common.loopingcall [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] task run outlasted interval by 0.30767199999999995 sec [ 647.727472] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453232, 'name': ReconfigVM_Task, 'duration_secs': 1.247759} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.734960] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Reconfigured VM instance instance-00000006 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 647.734960] env[65385]: DEBUG oslo_vmware.api [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453235, 'name': PowerOnVM_Task, 'duration_secs': 0.591254} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.734960] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453233, 'name': CreateVM_Task, 'duration_secs': 0.570882} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.735733] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4783c80-84bc-4adc-aac1-86fe055d76a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.738603] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 647.738812] env[65385]: INFO nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Took 10.51 seconds to spawn the instance on the hypervisor. [ 647.738993] env[65385]: DEBUG nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 647.739297] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 647.740815] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.741821] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86d51cd-ee92-4f04-8b48-d8122a3f5eca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.744855] env[65385]: WARNING neutronclient.v2_0.client [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.745152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.745313] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 647.745618] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 647.749040] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c83a4a6-9465-4731-a48d-18db22fd8ece {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.769721] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] 42b9edf6-1873-49c3-8074-8eef654ac371/42b9edf6-1873-49c3-8074-8eef654ac371.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 647.775600] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5e63ca8-d80e-4d1d-85ed-4ffc9e25ed70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.788795] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 647.788795] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b2257d-c90a-fb06-3d8d-291f9be9e662" [ 647.788795] env[65385]: _type = "Task" [ 647.788795] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.804631] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b2257d-c90a-fb06-3d8d-291f9be9e662, 'name': SearchDatastore_Task, 'duration_secs': 0.015312} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.804927] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 647.804927] env[65385]: value = "task-4453236" [ 647.804927] env[65385]: _type = "Task" [ 647.804927] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.805220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.805547] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 647.805827] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.805984] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 647.806183] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 647.806526] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fce03992-d81f-4100-a3c5-85f8eae8dcc0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.823010] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453236, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.836465] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 647.836740] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 647.840544] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4a1923b-6ec6-4db1-904f-5c1df7f1e4a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.848137] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 647.848137] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fafad-0617-71c6-e1bf-9ef335cc9db8" [ 647.848137] env[65385]: _type = "Task" [ 647.848137] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.858800] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fafad-0617-71c6-e1bf-9ef335cc9db8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.988147] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2445f8c4-8337-4839-a9de-d6cd981450b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.997099] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8bc2df-ebe7-48dd-a227-f3233358f47e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.034180] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.034180] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.040067] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.040570] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.048277] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0619437-9231-40d1-b852-bf2438a442e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.058526] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4931fdf7-fbcb-4e62-8cd7-76dbbca9e208 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.075458] env[65385]: DEBUG nova.compute.provider_tree [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.202281] env[65385]: INFO nova.compute.manager [-] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Took 2.13 seconds to deallocate network for instance. [ 648.306826] env[65385]: INFO nova.compute.manager [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Took 31.55 seconds to build instance. [ 648.323462] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453236, 'name': ReconfigVM_Task, 'duration_secs': 0.51238} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.324217] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Reconfigured VM instance instance-00000006 to attach disk [datastore1] 42b9edf6-1873-49c3-8074-8eef654ac371/42b9edf6-1873-49c3-8074-8eef654ac371.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 648.324217] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance '42b9edf6-1873-49c3-8074-8eef654ac371' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 648.364370] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fafad-0617-71c6-e1bf-9ef335cc9db8, 'name': SearchDatastore_Task, 'duration_secs': 0.012527} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.365615] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89dbfb62-11bb-448d-8130-b25c73af6394 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.373910] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 648.373910] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a89fe9-4a3b-99d3-580c-c8713aeff42f" [ 648.373910] env[65385]: _type = "Task" [ 648.373910] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.388782] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a89fe9-4a3b-99d3-580c-c8713aeff42f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.483981] env[65385]: WARNING neutronclient.v2_0.client [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.484725] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.485099] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.541169] env[65385]: WARNING neutronclient.v2_0.client [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.541839] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.542216] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.579051] env[65385]: DEBUG nova.scheduler.client.report [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 648.713098] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.740187] env[65385]: DEBUG nova.network.neutron [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Updated VIF entry in instance network info cache for port 4343bbc8-b6b2-4198-9f4c-d6da5ff242dc. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 648.740187] env[65385]: DEBUG nova.network.neutron [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Updating instance_info_cache with network_info: [{"id": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "address": "fa:16:3e:89:46:11", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4343bbc8-b6", "ovs_interfaceid": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 648.774864] env[65385]: DEBUG nova.network.neutron [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updated VIF entry in instance network info cache for port e37bf46a-308d-44a5-b573-64c3b7aa5a73. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 648.776267] env[65385]: DEBUG nova.network.neutron [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updating instance_info_cache with network_info: [{"id": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "address": "fa:16:3e:a7:7e:98", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape37bf46a-30", "ovs_interfaceid": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 648.817155] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cffe422b-ae97-41e9-aa64-0669bd72a698 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.263s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.832473] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8a2a28-a22c-4b08-84dd-b2cb416e5484 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.856136] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1a66da-1238-4408-9502-00051ea3ff46 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.877902] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance '42b9edf6-1873-49c3-8074-8eef654ac371' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 648.892362] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a89fe9-4a3b-99d3-580c-c8713aeff42f, 'name': SearchDatastore_Task, 'duration_secs': 0.014203} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.892723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.892997] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 75ba706f-2e72-4f84-b02f-db4381951e77/75ba706f-2e72-4f84-b02f-db4381951e77.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 648.893384] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e42e2d8e-7b39-47f9-940f-10d216c78c6a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.901913] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 648.901913] env[65385]: value = "task-4453237" [ 648.901913] env[65385]: _type = "Task" [ 648.901913] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.914796] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.085203] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.445s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 649.085756] env[65385]: DEBUG nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 649.089627] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.021s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.089967] env[65385]: DEBUG nova.objects.instance [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 649.243025] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Releasing lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 649.243025] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Received event network-vif-plugged-2206d06f-505b-4211-9823-29d900820228 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 649.243025] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Acquiring lock "eb9d0510-b453-4695-9e1d-731217b9f8ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.243025] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.243025] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 649.243938] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] No waiting events found dispatching network-vif-plugged-2206d06f-505b-4211-9823-29d900820228 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 649.243938] env[65385]: WARNING nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Received unexpected event network-vif-plugged-2206d06f-505b-4211-9823-29d900820228 for instance with vm_state building and task_state spawning. [ 649.243938] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Received event network-changed-2206d06f-505b-4211-9823-29d900820228 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 649.243938] env[65385]: DEBUG nova.compute.manager [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Refreshing instance network info cache due to event network-changed-2206d06f-505b-4211-9823-29d900820228. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 649.243938] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Acquiring lock "refresh_cache-eb9d0510-b453-4695-9e1d-731217b9f8ec" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.244292] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Acquired lock "refresh_cache-eb9d0510-b453-4695-9e1d-731217b9f8ec" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 649.244457] env[65385]: DEBUG nova.network.neutron [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Refreshing network info cache for port 2206d06f-505b-4211-9823-29d900820228 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 649.280792] env[65385]: DEBUG oslo_concurrency.lockutils [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Releasing lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 649.281154] env[65385]: DEBUG nova.compute.manager [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Received event network-vif-deleted-34c764b8-f45c-4237-9998-f49b871bb7d4 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 649.281402] env[65385]: DEBUG nova.compute.manager [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Received event network-changed-e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 649.281595] env[65385]: DEBUG nova.compute.manager [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Refreshing instance network info cache due to event network-changed-e37bf46a-308d-44a5-b573-64c3b7aa5a73. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 649.281840] env[65385]: DEBUG oslo_concurrency.lockutils [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Acquiring lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.283272] env[65385]: DEBUG oslo_concurrency.lockutils [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Acquired lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 649.283648] env[65385]: DEBUG nova.network.neutron [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Refreshing network info cache for port e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 649.390182] env[65385]: WARNING neutronclient.v2_0.client [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.390701] env[65385]: WARNING neutronclient.v2_0.client [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.415426] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453237, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.596514] env[65385]: DEBUG nova.compute.utils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 649.604599] env[65385]: DEBUG nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 649.604599] env[65385]: DEBUG nova.network.neutron [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 649.605645] env[65385]: WARNING neutronclient.v2_0.client [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.605645] env[65385]: WARNING neutronclient.v2_0.client [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.605876] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.606279] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.615564] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d85573-0f09-428e-98e1-7aeb83897fff tempest-ServersAdmin275Test-1725316109 tempest-ServersAdmin275Test-1725316109-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.526s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 649.617551] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.767s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.620575] env[65385]: INFO nova.compute.claims [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.747390] env[65385]: WARNING neutronclient.v2_0.client [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.748111] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.748800] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.787591] env[65385]: WARNING neutronclient.v2_0.client [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.788164] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.788497] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.916025] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453237, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.708196} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.916478] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 75ba706f-2e72-4f84-b02f-db4381951e77/75ba706f-2e72-4f84-b02f-db4381951e77.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 649.916610] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 649.916918] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-390ba247-9a55-4992-9473-7e6fb724c265 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.924928] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 649.924928] env[65385]: value = "task-4453238" [ 649.924928] env[65385]: _type = "Task" [ 649.924928] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.934694] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453238, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.102958] env[65385]: DEBUG nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 650.134308] env[65385]: DEBUG nova.network.neutron [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Port 03771ba6-a616-4715-afb9-19306b1b0903 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 650.247774] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.247774] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.343889] env[65385]: DEBUG nova.policy [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '464b0f7c8fe3426bacdd58396e49db51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d0195ab78e1455fab0eaf7ea55aea99', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 650.439382] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453238, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070911} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.439706] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 650.440639] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d843f819-4008-4b43-9bd4-f7d8ae5a84d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.465588] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] 75ba706f-2e72-4f84-b02f-db4381951e77/75ba706f-2e72-4f84-b02f-db4381951e77.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 650.469031] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4864ba54-6942-4061-9cfc-69a5a27905b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.487458] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 650.487458] env[65385]: value = "task-4453239" [ 650.487458] env[65385]: _type = "Task" [ 650.487458] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.496751] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.763096] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.763096] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.763096] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.763096] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.763096] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.763096] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.763872] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 650.763872] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.791642] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.792133] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.807913] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.808305] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.820331] env[65385]: DEBUG nova.network.neutron [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Successfully created port: 7454821e-9f2a-416b-a1ff-af59335a59a9 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 650.999564] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453239, 'name': ReconfigVM_Task, 'duration_secs': 0.309619} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.002526] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Reconfigured VM instance instance-00000015 to attach disk [datastore1] 75ba706f-2e72-4f84-b02f-db4381951e77/75ba706f-2e72-4f84-b02f-db4381951e77.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 651.004239] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6407f264-5f95-4f81-8d0d-4256a1cbad3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.010893] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 651.010893] env[65385]: value = "task-4453240" [ 651.010893] env[65385]: _type = "Task" [ 651.010893] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.023369] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453240, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.078489] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e31770-3d53-4290-ae1f-fe1d0d3f7b38 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.086999] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b854be1-a6b8-4627-9362-815f1ec492d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.122619] env[65385]: DEBUG nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 651.125422] env[65385]: WARNING neutronclient.v2_0.client [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.126211] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.126648] env[65385]: WARNING openstack [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.136731] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89335072-b6dc-443b-b6a6-d4f5d43e031c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.144861] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44fa4af-1103-4a37-914b-fd8f28d04204 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.176716] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.176963] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.177151] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 651.178659] env[65385]: DEBUG nova.compute.provider_tree [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.182249] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 651.182405] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 651.182565] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 651.182737] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 651.182880] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 651.183056] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 651.183274] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.183415] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 651.183572] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 651.183724] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 651.183887] env[65385]: DEBUG nova.virt.hardware [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 651.185057] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63ddfdf-2940-4a96-9ebf-bb85bfaca093 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.195687] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bec4a40-1888-480e-8945-c2e9b6395d3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.265220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.454183] env[65385]: WARNING neutronclient.v2_0.client [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.454914] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.455353] env[65385]: WARNING openstack [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.522041] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453240, 'name': Rename_Task, 'duration_secs': 0.164854} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.522568] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 651.523254] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a5d7cee-e7f3-4ace-96c4-a469afbf035c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.530328] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 651.530328] env[65385]: value = "task-4453241" [ 651.530328] env[65385]: _type = "Task" [ 651.530328] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.538661] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453241, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.689682] env[65385]: DEBUG nova.scheduler.client.report [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 652.043046] env[65385]: DEBUG oslo_vmware.api [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453241, 'name': PowerOnVM_Task, 'duration_secs': 0.482156} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.043046] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 652.043457] env[65385]: INFO nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Took 8.62 seconds to spawn the instance on the hypervisor. [ 652.043457] env[65385]: DEBUG nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 652.044681] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db21b814-d788-440c-b44f-09fe3d258f6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.315345] env[65385]: WARNING neutronclient.v2_0.client [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.315345] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 652.315345] env[65385]: DEBUG nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 652.315345] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.022s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 652.315345] env[65385]: INFO nova.compute.claims [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.475734] env[65385]: DEBUG nova.network.neutron [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Updated VIF entry in instance network info cache for port 2206d06f-505b-4211-9823-29d900820228. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 652.475734] env[65385]: DEBUG nova.network.neutron [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Updating instance_info_cache with network_info: [{"id": "2206d06f-505b-4211-9823-29d900820228", "address": "fa:16:3e:81:0f:eb", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2206d06f-50", "ovs_interfaceid": "2206d06f-505b-4211-9823-29d900820228", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 652.536028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.536028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.536165] env[65385]: DEBUG nova.network.neutron [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 652.570873] env[65385]: INFO nova.compute.manager [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Took 31.07 seconds to build instance. [ 652.673097] env[65385]: DEBUG nova.network.neutron [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Successfully updated port: 7454821e-9f2a-416b-a1ff-af59335a59a9 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 652.713323] env[65385]: DEBUG nova.network.neutron [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updated VIF entry in instance network info cache for port e37bf46a-308d-44a5-b573-64c3b7aa5a73. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 652.714767] env[65385]: DEBUG nova.network.neutron [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updating instance_info_cache with network_info: [{"id": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "address": "fa:16:3e:a7:7e:98", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape37bf46a-30", "ovs_interfaceid": "e37bf46a-308d-44a5-b573-64c3b7aa5a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 652.718607] env[65385]: DEBUG nova.compute.utils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 652.723983] env[65385]: DEBUG nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 652.723983] env[65385]: DEBUG nova.network.neutron [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 652.724688] env[65385]: WARNING neutronclient.v2_0.client [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.725024] env[65385]: WARNING neutronclient.v2_0.client [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.726066] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.726066] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.782899] env[65385]: DEBUG nova.policy [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '645e587c101449ed8a42b465352640c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76d02fba23ea453a9621e58d6b3703a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 652.979293] env[65385]: DEBUG oslo_concurrency.lockutils [req-93147bdc-6d76-4f96-9a80-9de313c5190e req-5ae7564f-778b-4f46-840c-245e8ce85308 service nova] Releasing lock "refresh_cache-eb9d0510-b453-4695-9e1d-731217b9f8ec" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.039610] env[65385]: WARNING neutronclient.v2_0.client [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 653.039957] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.040207] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.073438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93611b2d-c3c0-4150-9970-82def48867a9 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "75ba706f-2e72-4f84-b02f-db4381951e77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.579s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.177806] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "refresh_cache-69fdd5df-a9f2-486f-8a79-87c034366083" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.177967] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired lock "refresh_cache-69fdd5df-a9f2-486f-8a79-87c034366083" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 653.178177] env[65385]: DEBUG nova.network.neutron [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 653.226362] env[65385]: DEBUG oslo_concurrency.lockutils [req-caa2d709-da4d-443d-8ffd-5c9c808b62d4 req-3355ab2d-9cd1-451a-9a47-0ac9374cfbe2 service nova] Releasing lock "refresh_cache-0115dde8-b0aa-4960-a452-ced6587bc567" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.227834] env[65385]: DEBUG nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 653.326239] env[65385]: DEBUG nova.network.neutron [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Successfully created port: e6913ace-1d77-409d-8108-5f01780b3051 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 653.489877] env[65385]: DEBUG nova.compute.manager [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Received event network-vif-plugged-d59b902f-9fb1-4d56-96f9-6301024412fa {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 653.490184] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Acquiring lock "75ba706f-2e72-4f84-b02f-db4381951e77-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.490721] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Lock "75ba706f-2e72-4f84-b02f-db4381951e77-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.491035] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Lock "75ba706f-2e72-4f84-b02f-db4381951e77-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.491843] env[65385]: DEBUG nova.compute.manager [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] No waiting events found dispatching network-vif-plugged-d59b902f-9fb1-4d56-96f9-6301024412fa {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 653.491900] env[65385]: WARNING nova.compute.manager [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Received unexpected event network-vif-plugged-d59b902f-9fb1-4d56-96f9-6301024412fa for instance with vm_state active and task_state None. [ 653.492509] env[65385]: DEBUG nova.compute.manager [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Received event network-changed-d59b902f-9fb1-4d56-96f9-6301024412fa {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 653.492509] env[65385]: DEBUG nova.compute.manager [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Refreshing instance network info cache due to event network-changed-d59b902f-9fb1-4d56-96f9-6301024412fa. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 653.493354] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Acquiring lock "refresh_cache-75ba706f-2e72-4f84-b02f-db4381951e77" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.493354] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Acquired lock "refresh_cache-75ba706f-2e72-4f84-b02f-db4381951e77" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 653.493354] env[65385]: DEBUG nova.network.neutron [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Refreshing network info cache for port d59b902f-9fb1-4d56-96f9-6301024412fa {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 653.675271] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85b0363-6c4e-4e59-b478-c1c1909bdab6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.681784] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.682398] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.692476] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607b8b68-4f86-4aa6-b6b2-f993abeafeab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.729556] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd60a16-6a28-4dac-9864-c6b3afdb22df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.743302] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffb5612-9c19-4243-9e6e-a30bab8d58e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.762394] env[65385]: DEBUG nova.compute.provider_tree [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.997264] env[65385]: WARNING neutronclient.v2_0.client [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 653.997936] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.998341] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.211337] env[65385]: DEBUG nova.compute.manager [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Received event network-changed-7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 654.211663] env[65385]: DEBUG nova.compute.manager [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Refreshing instance network info cache due to event network-changed-7d3f8bf5-22d2-41b0-9743-0616daaef615. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 654.211764] env[65385]: DEBUG oslo_concurrency.lockutils [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Acquiring lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.212056] env[65385]: DEBUG oslo_concurrency.lockutils [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Acquired lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.212056] env[65385]: DEBUG nova.network.neutron [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Refreshing network info cache for port 7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 654.239301] env[65385]: DEBUG nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 654.269025] env[65385]: DEBUG nova.scheduler.client.report [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 654.282598] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 654.283426] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 654.283426] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 654.283646] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 654.283968] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 654.284934] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 654.284934] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.284934] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 654.284934] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 654.285188] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 654.285188] env[65385]: DEBUG nova.virt.hardware [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 654.286023] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57018b33-0ee6-4070-ba44-11a0047761f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.296487] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8817c07-88f1-4e60-9b5f-06e0507d4388 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.714754] env[65385]: WARNING neutronclient.v2_0.client [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.715556] env[65385]: WARNING openstack [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.715947] env[65385]: WARNING openstack [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.777014] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 654.777572] env[65385]: DEBUG nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 654.783050] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.661s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.783704] env[65385]: INFO nova.compute.claims [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.906284] env[65385]: DEBUG nova.network.neutron [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 654.923047] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.923640] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.005271] env[65385]: DEBUG nova.network.neutron [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Successfully updated port: e6913ace-1d77-409d-8108-5f01780b3051 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 655.049892] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.050330] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.290564] env[65385]: DEBUG nova.compute.utils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 655.297277] env[65385]: DEBUG nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 655.297550] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 655.298064] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.298472] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.299265] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.299754] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.310428] env[65385]: WARNING neutronclient.v2_0.client [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.311112] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.311496] env[65385]: WARNING openstack [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.325046] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.325548] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.338873] env[65385]: WARNING openstack [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.339528] env[65385]: WARNING openstack [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.406662] env[65385]: WARNING neutronclient.v2_0.client [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.408074] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.408475] env[65385]: WARNING openstack [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.460400] env[65385]: DEBUG nova.policy [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01c9e45d7e6c43eb920276c99d5a2e9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21132a4099d84aaa89698915f7e91c01', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 655.508670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "refresh_cache-de8632e0-bf38-440d-b6a3-895efeef122d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.508865] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquired lock "refresh_cache-de8632e0-bf38-440d-b6a3-895efeef122d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.509029] env[65385]: DEBUG nova.network.neutron [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 655.777464] env[65385]: WARNING neutronclient.v2_0.client [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.778531] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.778531] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.789229] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cadc63-4843-477a-8327-b00a103ecafd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.797110] env[65385]: DEBUG nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 655.800521] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418453cc-2f11-4686-9dd9-11a7bcd0620d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.837447] env[65385]: DEBUG nova.network.neutron [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance_info_cache with network_info: [{"id": "03771ba6-a616-4715-afb9-19306b1b0903", "address": "fa:16:3e:2b:4a:04", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03771ba6-a6", "ovs_interfaceid": "03771ba6-a616-4715-afb9-19306b1b0903", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 655.839174] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35320ad6-0eb5-4a8d-b96f-c20b902b3d05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.843930] env[65385]: WARNING neutronclient.v2_0.client [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.844945] env[65385]: WARNING openstack [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.845325] env[65385]: WARNING openstack [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.857971] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Successfully created port: b89f2175-8529-42c4-9a78-6e7116ad7293 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 655.860566] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470fcaff-65c8-4c6b-9854-8358f437a1ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.876647] env[65385]: DEBUG nova.compute.provider_tree [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.910957] env[65385]: DEBUG nova.network.neutron [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Updating instance_info_cache with network_info: [{"id": "7454821e-9f2a-416b-a1ff-af59335a59a9", "address": "fa:16:3e:ee:8e:02", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7454821e-9f", "ovs_interfaceid": "7454821e-9f2a-416b-a1ff-af59335a59a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.011677] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.011981] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.346359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.365052] env[65385]: DEBUG nova.network.neutron [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 656.380037] env[65385]: DEBUG nova.scheduler.client.report [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.412411] env[65385]: DEBUG nova.network.neutron [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Updated VIF entry in instance network info cache for port d59b902f-9fb1-4d56-96f9-6301024412fa. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 656.412913] env[65385]: DEBUG nova.network.neutron [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Updating instance_info_cache with network_info: [{"id": "d59b902f-9fb1-4d56-96f9-6301024412fa", "address": "fa:16:3e:f2:b1:5a", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd59b902f-9f", "ovs_interfaceid": "d59b902f-9fb1-4d56-96f9-6301024412fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.414649] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Releasing lock "refresh_cache-69fdd5df-a9f2-486f-8a79-87c034366083" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.415040] env[65385]: DEBUG nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Instance network_info: |[{"id": "7454821e-9f2a-416b-a1ff-af59335a59a9", "address": "fa:16:3e:ee:8e:02", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7454821e-9f", "ovs_interfaceid": "7454821e-9f2a-416b-a1ff-af59335a59a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 656.417312] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Successfully created port: f23a0b59-d6a0-416e-a41c-d0f1f12672ee {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 656.419614] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:8e:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7454821e-9f2a-416b-a1ff-af59335a59a9', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 656.427514] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 656.429057] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 656.429194] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de437d91-010e-46a4-bb61-c52d0dc63a59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.446877] env[65385]: DEBUG nova.network.neutron [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updated VIF entry in instance network info cache for port 7d3f8bf5-22d2-41b0-9743-0616daaef615. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 656.447207] env[65385]: DEBUG nova.network.neutron [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updating instance_info_cache with network_info: [{"id": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "address": "fa:16:3e:f9:84:04", "network": {"id": "9ed70dba-8022-422b-8af1-9e125786bbe0", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-823001245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73c43e9322314241811db1a37e18f9b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3f8bf5-22", "ovs_interfaceid": "7d3f8bf5-22d2-41b0-9743-0616daaef615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.456247] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 656.456247] env[65385]: value = "task-4453242" [ 656.456247] env[65385]: _type = "Task" [ 656.456247] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.467266] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453242, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.483169] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.483681] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.761829] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Successfully created port: 8c68958e-7ea9-4a06-b35a-05c48985738b {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 656.814145] env[65385]: DEBUG nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 656.846490] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 656.846490] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 656.846616] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 656.846712] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 656.847257] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 656.847428] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 656.847937] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.848125] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 656.848295] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 656.848484] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 656.848665] env[65385]: DEBUG nova.virt.hardware [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 656.850046] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eacf353-430e-46aa-93cd-62cc283771b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.871083] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c38233-3ef2-4fbd-98cb-a1c823d539a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.892272] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.893034] env[65385]: DEBUG nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 656.902077] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.527s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.902553] env[65385]: DEBUG nova.objects.instance [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'resources' on Instance uuid f77d861a-ae66-44c3-901d-e499926d6f5b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 656.904485] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2e7745-4836-4e04-9f48-35aee9747e4d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.926217] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Releasing lock "refresh_cache-75ba706f-2e72-4f84-b02f-db4381951e77" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.926394] env[65385]: DEBUG nova.compute.manager [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Received event network-changed-97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 656.926615] env[65385]: DEBUG nova.compute.manager [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Refreshing instance network info cache due to event network-changed-97c1d2c0-12b5-428b-85b4-81b8305d87f9. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 656.926818] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Acquiring lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.926987] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Acquired lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 656.927159] env[65385]: DEBUG nova.network.neutron [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Refreshing network info cache for port 97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 656.929598] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45d168a-0ec7-4393-a4e3-b7727e01998a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.938793] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance '42b9edf6-1873-49c3-8074-8eef654ac371' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 656.951802] env[65385]: DEBUG oslo_concurrency.lockutils [req-4771a3e6-b6ce-4102-94da-4a52ddadbbb1 req-438f7f2c-9eb8-414d-976b-e0cb5e36da07 service nova] Releasing lock "refresh_cache-bf79fbcd-e943-4ff3-bcf8-6a2817df8292" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.968000] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453242, 'name': CreateVM_Task, 'duration_secs': 0.391433} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.968000] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 656.968000] env[65385]: WARNING neutronclient.v2_0.client [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 656.968275] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.968403] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 656.968704] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 656.969272] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31c7d8c0-7caf-421d-a4a3-81c6443db1b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.975110] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 656.975110] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521f85f1-d1ef-aa2f-40bb-43bfeea7517a" [ 656.975110] env[65385]: _type = "Task" [ 656.975110] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.984402] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521f85f1-d1ef-aa2f-40bb-43bfeea7517a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.134388] env[65385]: WARNING neutronclient.v2_0.client [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.135670] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.135874] env[65385]: WARNING openstack [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.297109] env[65385]: DEBUG nova.network.neutron [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Updating instance_info_cache with network_info: [{"id": "e6913ace-1d77-409d-8108-5f01780b3051", "address": "fa:16:3e:5f:ed:61", "network": {"id": "5cadf630-343b-40b7-a126-ec9e712d5c7f", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1639848170-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76d02fba23ea453a9621e58d6b3703a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62237242-7ce2-4664-a1c5-6783b516b507", "external-id": "nsx-vlan-transportzone-295", "segmentation_id": 295, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6913ace-1d", "ovs_interfaceid": "e6913ace-1d77-409d-8108-5f01780b3051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 657.410447] env[65385]: DEBUG nova.compute.utils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 657.416186] env[65385]: DEBUG nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 657.416419] env[65385]: DEBUG nova.network.neutron [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 657.416879] env[65385]: WARNING neutronclient.v2_0.client [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.417028] env[65385]: WARNING neutronclient.v2_0.client [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.417597] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.417948] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.433310] env[65385]: WARNING neutronclient.v2_0.client [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.433935] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.434319] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.445486] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 657.445955] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-abb1bf98-266c-4986-b787-9d696feb4622 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.453386] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 657.453386] env[65385]: value = "task-4453243" [ 657.453386] env[65385]: _type = "Task" [ 657.453386] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.464660] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453243, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.486111] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521f85f1-d1ef-aa2f-40bb-43bfeea7517a, 'name': SearchDatastore_Task, 'duration_secs': 0.010263} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.486533] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.486688] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 657.487083] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.487164] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.487320] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 657.490276] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-241c67c0-2166-43ec-b534-ee8a507b972f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.499876] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 657.500077] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 657.500837] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85d001d7-d565-4dac-9eba-2555204704d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.509633] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 657.509633] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5208aa83-410b-644a-32d2-3acf81064554" [ 657.509633] env[65385]: _type = "Task" [ 657.509633] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.519096] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5208aa83-410b-644a-32d2-3acf81064554, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.575502] env[65385]: DEBUG nova.policy [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac74a729b62042eda193030d716719e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1598307c304d46a884a88b3eb5f5d7dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 657.773076] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18c4322-84cf-4bb1-a33d-65117007fd94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.783745] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bedd2a-cd86-48e4-bfd4-4064de67c3a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.818691] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Releasing lock "refresh_cache-de8632e0-bf38-440d-b6a3-895efeef122d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.819450] env[65385]: DEBUG nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Instance network_info: |[{"id": "e6913ace-1d77-409d-8108-5f01780b3051", "address": "fa:16:3e:5f:ed:61", "network": {"id": "5cadf630-343b-40b7-a126-ec9e712d5c7f", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1639848170-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76d02fba23ea453a9621e58d6b3703a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62237242-7ce2-4664-a1c5-6783b516b507", "external-id": "nsx-vlan-transportzone-295", "segmentation_id": 295, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6913ace-1d", "ovs_interfaceid": "e6913ace-1d77-409d-8108-5f01780b3051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 657.820374] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:ed:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62237242-7ce2-4664-a1c5-6783b516b507', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e6913ace-1d77-409d-8108-5f01780b3051', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 657.829337] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Creating folder: Project (76d02fba23ea453a9621e58d6b3703a0). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 657.830607] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4736fa5b-f19f-44fa-96fa-822c1c3846f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.834090] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c3a6bd68-ae43-48ab-afbb-a6867227d6dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.847188] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362e2202-cce6-42e1-9c88-e8d77ddae275 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.850465] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Created folder: Project (76d02fba23ea453a9621e58d6b3703a0) in parent group-v870881. [ 657.850812] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Creating folder: Instances. Parent ref: group-v870942. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 657.851624] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1dc9e7d5-5a60-4d24-99f3-2958bb896eee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.862157] env[65385]: DEBUG nova.compute.provider_tree [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.866127] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Created folder: Instances in parent group-v870942. [ 657.866728] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 657.867094] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 657.867506] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-792d0d03-7a8f-48df-9115-304f0cb9134a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.890418] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 657.890418] env[65385]: value = "task-4453246" [ 657.890418] env[65385]: _type = "Task" [ 657.890418] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.900367] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453246, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.916818] env[65385]: DEBUG nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 657.966246] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453243, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.023044] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5208aa83-410b-644a-32d2-3acf81064554, 'name': SearchDatastore_Task, 'duration_secs': 0.017082} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.023960] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ba6c88a-1a41-437d-a626-f48ee83ae3f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.031678] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 658.031678] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52474b4b-ec0e-9933-85a2-e72f3fa747a5" [ 658.031678] env[65385]: _type = "Task" [ 658.031678] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.042599] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52474b4b-ec0e-9933-85a2-e72f3fa747a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.278746] env[65385]: DEBUG nova.network.neutron [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Successfully created port: 18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 658.365874] env[65385]: DEBUG nova.scheduler.client.report [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 658.402966] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453246, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.437090] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.437507] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.467719] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453243, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.545825] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52474b4b-ec0e-9933-85a2-e72f3fa747a5, 'name': SearchDatastore_Task, 'duration_secs': 0.014586} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.545825] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.545825] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 69fdd5df-a9f2-486f-8a79-87c034366083/69fdd5df-a9f2-486f-8a79-87c034366083.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 658.545825] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61d1ad63-ba04-4be4-a8f6-9b47a53e54b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.554312] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 658.554312] env[65385]: value = "task-4453247" [ 658.554312] env[65385]: _type = "Task" [ 658.554312] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.563423] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453247, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.610323] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Successfully updated port: b89f2175-8529-42c4-9a78-6e7116ad7293 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 658.876343] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 658.879493] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.318s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 658.889033] env[65385]: INFO nova.compute.claims [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.904471] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453246, 'name': CreateVM_Task, 'duration_secs': 0.526881} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.904616] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 658.905253] env[65385]: WARNING neutronclient.v2_0.client [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.905713] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.906228] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 658.906640] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 658.906677] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c61e523d-9abd-409e-be89-83bdcf3025d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.916933] env[65385]: INFO nova.scheduler.client.report [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted allocations for instance f77d861a-ae66-44c3-901d-e499926d6f5b [ 658.926848] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 658.927761] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 658.929034] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 658.929034] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520a1810-20a2-b5f5-dde1-5bc17a3b3eca" [ 658.929034] env[65385]: _type = "Task" [ 658.929034] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.932548] env[65385]: DEBUG nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 658.944601] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520a1810-20a2-b5f5-dde1-5bc17a3b3eca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.969975] env[65385]: DEBUG oslo_vmware.api [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453243, 'name': PowerOnVM_Task, 'duration_secs': 1.015103} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.972462] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 658.972978] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 658.973241] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 658.973483] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 658.973756] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 658.973857] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 658.974097] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.974339] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 658.974570] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 658.974781] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 658.975177] env[65385]: DEBUG nova.virt.hardware [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 658.975592] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 658.975773] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa0e22a-4eea-47c1-95e4-add5b880df13 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance '42b9edf6-1873-49c3-8074-8eef654ac371' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 658.981575] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1321400e-c573-44d0-a92f-9c79e3fdaf89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.991753] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6f6c7f-44d5-49cb-89d8-f9163821b061 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.066767] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453247, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.372244] env[65385]: WARNING neutronclient.v2_0.client [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.372896] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.373252] env[65385]: WARNING openstack [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.434900] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9bd63f3-292f-4caa-840d-ab0e70d656b3 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "f77d861a-ae66-44c3-901d-e499926d6f5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.732s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.436938] env[65385]: DEBUG nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 659.453486] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520a1810-20a2-b5f5-dde1-5bc17a3b3eca, 'name': SearchDatastore_Task, 'duration_secs': 0.059943} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.454414] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.454688] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 659.454929] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.455091] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.455301] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 659.456035] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98c5366c-71e6-417f-9ede-929d79cfc3ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.469573] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 659.469791] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 659.471046] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-238f4ed5-8874-4c42-a3cb-f41d93df3309 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.478631] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 659.478631] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fb9bb-6564-259a-07a7-a6ded21f2c98" [ 659.478631] env[65385]: _type = "Task" [ 659.478631] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.491191] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fb9bb-6564-259a-07a7-a6ded21f2c98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.565808] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453247, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670408} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.566072] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 69fdd5df-a9f2-486f-8a79-87c034366083/69fdd5df-a9f2-486f-8a79-87c034366083.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 659.566275] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 659.566569] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dceab96d-22af-4b7c-ad4c-906989084d14 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.573934] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 659.573934] env[65385]: value = "task-4453248" [ 659.573934] env[65385]: _type = "Task" [ 659.573934] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.583720] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453248, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.966847] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 659.995819] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fb9bb-6564-259a-07a7-a6ded21f2c98, 'name': SearchDatastore_Task, 'duration_secs': 0.060788} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.999808] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-191049bd-da2c-40c5-b2f6-87e0f5d0ae6c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.008063] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 660.008063] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b5b3ea-5795-0b09-4820-5923b08eacd3" [ 660.008063] env[65385]: _type = "Task" [ 660.008063] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.014182] env[65385]: DEBUG nova.network.neutron [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Successfully updated port: 18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 660.025482] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b5b3ea-5795-0b09-4820-5923b08eacd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.085325] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453248, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.207251} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.085638] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 660.086451] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fbc8e79-cdea-4ef2-aaa6-9921cfebddaf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.115035] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 69fdd5df-a9f2-486f-8a79-87c034366083/69fdd5df-a9f2-486f-8a79-87c034366083.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 660.118560] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dc0dc0f-4281-4c33-85a7-1ad90ea5d2e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.141917] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 660.141917] env[65385]: value = "task-4453249" [ 660.141917] env[65385]: _type = "Task" [ 660.141917] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.155405] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453249, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.341781] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e60eee-d2d0-46fe-99f0-c8b253483ea1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.351091] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714d2862-f533-43a9-a4cb-06e2aaf63f9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.388761] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e873a59e-9187-46c2-8b42-3a013b248292 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.398059] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d41b20-fe60-455c-991d-3f5479b3a736 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.412352] env[65385]: DEBUG nova.compute.provider_tree [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 660.523047] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b5b3ea-5795-0b09-4820-5923b08eacd3, 'name': SearchDatastore_Task, 'duration_secs': 0.013344} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.523441] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.523535] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] de8632e0-bf38-440d-b6a3-895efeef122d/de8632e0-bf38-440d-b6a3-895efeef122d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 660.523792] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34b42a14-6275-49fd-a3c4-a6cfe2d9c214 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.527473] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.527473] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.527690] env[65385]: DEBUG nova.network.neutron [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 660.533255] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 660.533255] env[65385]: value = "task-4453250" [ 660.533255] env[65385]: _type = "Task" [ 660.533255] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.544450] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453250, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.654461] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453249, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.937999] env[65385]: ERROR nova.scheduler.client.report [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [req-85fbca0f-0146-4b89-8802-7dcdc9b68d53] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-85fbca0f-0146-4b89-8802-7dcdc9b68d53"}]} [ 660.955489] env[65385]: DEBUG nova.scheduler.client.report [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 660.971137] env[65385]: DEBUG nova.scheduler.client.report [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 660.971431] env[65385]: DEBUG nova.compute.provider_tree [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 660.985291] env[65385]: DEBUG nova.scheduler.client.report [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 661.005931] env[65385]: DEBUG nova.scheduler.client.report [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 661.009453] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Successfully updated port: f23a0b59-d6a0-416e-a41c-d0f1f12672ee {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 661.032647] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.033301] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.048022] env[65385]: DEBUG nova.network.neutron [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Updated VIF entry in instance network info cache for port 97c1d2c0-12b5-428b-85b4-81b8305d87f9. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 661.048132] env[65385]: DEBUG nova.network.neutron [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Updating instance_info_cache with network_info: [{"id": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "address": "fa:16:3e:3a:2c:78", "network": {"id": "a2c91019-c13d-4581-aa0b-05553afc8f57", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1163204339-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f858126fa23d43cbafd1b677206f28ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97c1d2c0-12", "ovs_interfaceid": "97c1d2c0-12b5-428b-85b4-81b8305d87f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 661.065437] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453250, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.159631] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453249, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.193495] env[65385]: DEBUG nova.network.neutron [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 661.337011] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.337394] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.421567] env[65385]: WARNING neutronclient.v2_0.client [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.422437] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.422790] env[65385]: WARNING openstack [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.479023] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9525814-9322-4fb4-824f-53a0555519e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.489631] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998ff3a5-9e38-43b2-9ecd-9f9068ac3729 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.524497] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113e3ce7-4cd2-448f-bf64-b5dbae1d4d62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.537184] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4810521-6965-4292-be88-47e109a3e535 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.553024] env[65385]: DEBUG nova.compute.provider_tree [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 661.557978] env[65385]: DEBUG oslo_concurrency.lockutils [req-ee23e790-e274-4d61-b8db-d599aa6a0e47 req-f875ef55-52ee-4c71-8c83-d20754a0eb1b service nova] Releasing lock "refresh_cache-59450963-d83f-46e8-8c13-05d4f1818c64" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.564958] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453250, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.872191} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.566092] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] de8632e0-bf38-440d-b6a3-895efeef122d/de8632e0-bf38-440d-b6a3-895efeef122d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 661.566092] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 661.566384] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7899bf68-beb9-4e89-a696-d1f93fa18acb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.575898] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 661.575898] env[65385]: value = "task-4453251" [ 661.575898] env[65385]: _type = "Task" [ 661.575898] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.588858] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453251, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.640142] env[65385]: DEBUG nova.compute.manager [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Received event network-vif-plugged-e6913ace-1d77-409d-8108-5f01780b3051 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 661.640142] env[65385]: DEBUG oslo_concurrency.lockutils [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Acquiring lock "de8632e0-bf38-440d-b6a3-895efeef122d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.640142] env[65385]: DEBUG oslo_concurrency.lockutils [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Lock "de8632e0-bf38-440d-b6a3-895efeef122d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.640142] env[65385]: DEBUG oslo_concurrency.lockutils [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Lock "de8632e0-bf38-440d-b6a3-895efeef122d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.640142] env[65385]: DEBUG nova.compute.manager [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] No waiting events found dispatching network-vif-plugged-e6913ace-1d77-409d-8108-5f01780b3051 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 661.640376] env[65385]: WARNING nova.compute.manager [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Received unexpected event network-vif-plugged-e6913ace-1d77-409d-8108-5f01780b3051 for instance with vm_state building and task_state spawning. [ 661.640620] env[65385]: DEBUG nova.compute.manager [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Received event network-changed-e6913ace-1d77-409d-8108-5f01780b3051 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 661.640976] env[65385]: DEBUG nova.compute.manager [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Refreshing instance network info cache due to event network-changed-e6913ace-1d77-409d-8108-5f01780b3051. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 661.641358] env[65385]: DEBUG oslo_concurrency.lockutils [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Acquiring lock "refresh_cache-de8632e0-bf38-440d-b6a3-895efeef122d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.641649] env[65385]: DEBUG oslo_concurrency.lockutils [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Acquired lock "refresh_cache-de8632e0-bf38-440d-b6a3-895efeef122d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.642245] env[65385]: DEBUG nova.network.neutron [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Refreshing network info cache for port e6913ace-1d77-409d-8108-5f01780b3051 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 661.658059] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453249, 'name': ReconfigVM_Task, 'duration_secs': 1.510797} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.658599] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 69fdd5df-a9f2-486f-8a79-87c034366083/69fdd5df-a9f2-486f-8a79-87c034366083.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 661.659579] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c07565a7-4da6-46cd-8810-a7150b0436dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.668038] env[65385]: DEBUG nova.compute.manager [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Received event network-vif-plugged-7454821e-9f2a-416b-a1ff-af59335a59a9 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 661.668038] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Acquiring lock "69fdd5df-a9f2-486f-8a79-87c034366083-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.670748] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Lock "69fdd5df-a9f2-486f-8a79-87c034366083-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.670748] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Lock "69fdd5df-a9f2-486f-8a79-87c034366083-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.670748] env[65385]: DEBUG nova.compute.manager [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] No waiting events found dispatching network-vif-plugged-7454821e-9f2a-416b-a1ff-af59335a59a9 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 661.670748] env[65385]: WARNING nova.compute.manager [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Received unexpected event network-vif-plugged-7454821e-9f2a-416b-a1ff-af59335a59a9 for instance with vm_state building and task_state spawning. [ 661.670748] env[65385]: DEBUG nova.compute.manager [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Received event network-changed-7454821e-9f2a-416b-a1ff-af59335a59a9 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 661.671064] env[65385]: DEBUG nova.compute.manager [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Refreshing instance network info cache due to event network-changed-7454821e-9f2a-416b-a1ff-af59335a59a9. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 661.671064] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Acquiring lock "refresh_cache-69fdd5df-a9f2-486f-8a79-87c034366083" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.671064] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Acquired lock "refresh_cache-69fdd5df-a9f2-486f-8a79-87c034366083" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.671064] env[65385]: DEBUG nova.network.neutron [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Refreshing network info cache for port 7454821e-9f2a-416b-a1ff-af59335a59a9 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 661.674418] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 661.674418] env[65385]: value = "task-4453252" [ 661.674418] env[65385]: _type = "Task" [ 661.674418] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.688307] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453252, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.088293] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453251, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098032} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.089046] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 662.091158] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349dfcb1-ddea-4d7f-aec6-bc9bbe65104a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.095666] env[65385]: DEBUG nova.scheduler.client.report [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 36 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 662.096930] env[65385]: DEBUG nova.compute.provider_tree [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 36 to 37 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 662.097433] env[65385]: DEBUG nova.compute.provider_tree [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 662.127058] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] de8632e0-bf38-440d-b6a3-895efeef122d/de8632e0-bf38-440d-b6a3-895efeef122d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 662.128570] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a6a4703-2670-442b-ac06-31e41daa6956 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.146389] env[65385]: WARNING neutronclient.v2_0.client [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.146941] env[65385]: WARNING openstack [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.147323] env[65385]: WARNING openstack [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.164186] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 662.164186] env[65385]: value = "task-4453253" [ 662.164186] env[65385]: _type = "Task" [ 662.164186] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.174415] env[65385]: WARNING neutronclient.v2_0.client [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.174967] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.175369] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.182452] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.193695] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453252, 'name': Rename_Task, 'duration_secs': 0.472038} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.196308] env[65385]: DEBUG nova.network.neutron [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 662.197607] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 662.198378] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5bc95cc-96dd-4139-b63e-f2184b7476eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.206014] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 662.206014] env[65385]: value = "task-4453254" [ 662.206014] env[65385]: _type = "Task" [ 662.206014] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.217903] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453254, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.526023] env[65385]: WARNING openstack [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.526023] env[65385]: WARNING openstack [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.575858] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.576301] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.628880] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.749s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.631173] env[65385]: DEBUG nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 662.634916] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.308s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.635095] env[65385]: DEBUG nova.objects.instance [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lazy-loading 'resources' on Instance uuid c5b85574-df14-4a7a-ada0-c9668bf18a33 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 662.677670] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453253, 'name': ReconfigVM_Task, 'duration_secs': 0.376404} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.677938] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Reconfigured VM instance instance-00000017 to attach disk [datastore2] de8632e0-bf38-440d-b6a3-895efeef122d/de8632e0-bf38-440d-b6a3-895efeef122d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 662.678583] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65a7108f-5c62-421b-a8c0-f8533f12d741 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.686325] env[65385]: WARNING neutronclient.v2_0.client [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.686943] env[65385]: WARNING openstack [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.687292] env[65385]: WARNING openstack [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.698401] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 662.698401] env[65385]: value = "task-4453255" [ 662.698401] env[65385]: _type = "Task" [ 662.698401] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.698932] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 662.699278] env[65385]: DEBUG nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Instance network_info: |[{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 662.702883] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:d7:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9875d38f-76e2-416c-bfb7-f18a22b0d8ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18c18d61-0493-40e3-8883-c90faabf147b', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 662.711169] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating folder: Project (1598307c304d46a884a88b3eb5f5d7dc). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 662.712791] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5bd69b19-af29-4b0a-b520-99aa973d77d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.723225] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453255, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.730137] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453254, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.732371] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Created folder: Project (1598307c304d46a884a88b3eb5f5d7dc) in parent group-v870881. [ 662.732575] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating folder: Instances. Parent ref: group-v870945. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 662.732852] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16061b27-df37-44ce-b1d7-3bf8bf554684 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.744814] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Created folder: Instances in parent group-v870945. [ 662.744918] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 662.745163] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 662.745387] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70135e2d-4ed7-49ec-b560-aa8685c15682 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.768868] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 662.768868] env[65385]: value = "task-4453258" [ 662.768868] env[65385]: _type = "Task" [ 662.768868] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.778693] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453258, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.784462] env[65385]: WARNING neutronclient.v2_0.client [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.785173] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.785573] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.934785] env[65385]: DEBUG nova.network.neutron [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Updated VIF entry in instance network info cache for port e6913ace-1d77-409d-8108-5f01780b3051. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 662.935166] env[65385]: DEBUG nova.network.neutron [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Updating instance_info_cache with network_info: [{"id": "e6913ace-1d77-409d-8108-5f01780b3051", "address": "fa:16:3e:5f:ed:61", "network": {"id": "5cadf630-343b-40b7-a126-ec9e712d5c7f", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1639848170-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76d02fba23ea453a9621e58d6b3703a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62237242-7ce2-4664-a1c5-6783b516b507", "external-id": "nsx-vlan-transportzone-295", "segmentation_id": 295, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6913ace-1d", "ovs_interfaceid": "e6913ace-1d77-409d-8108-5f01780b3051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.038630] env[65385]: DEBUG nova.network.neutron [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Updated VIF entry in instance network info cache for port 7454821e-9f2a-416b-a1ff-af59335a59a9. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 663.039000] env[65385]: DEBUG nova.network.neutron [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Updating instance_info_cache with network_info: [{"id": "7454821e-9f2a-416b-a1ff-af59335a59a9", "address": "fa:16:3e:ee:8e:02", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7454821e-9f", "ovs_interfaceid": "7454821e-9f2a-416b-a1ff-af59335a59a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.140605] env[65385]: DEBUG nova.compute.utils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 663.143627] env[65385]: DEBUG nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 663.143859] env[65385]: DEBUG nova.network.neutron [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 663.144216] env[65385]: WARNING neutronclient.v2_0.client [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.144584] env[65385]: WARNING neutronclient.v2_0.client [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.145844] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.147309] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.209807] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453255, 'name': Rename_Task, 'duration_secs': 0.162361} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.214462] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 663.215074] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f8c2a61-a7d6-4b57-959f-e006e8d65985 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.221113] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Successfully updated port: 8c68958e-7ea9-4a06-b35a-05c48985738b {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 663.236166] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453254, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.240906] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 663.240906] env[65385]: value = "task-4453259" [ 663.240906] env[65385]: _type = "Task" [ 663.240906] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.254717] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453259, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.256268] env[65385]: DEBUG nova.policy [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '405200cf79e44e929194fc1989da78a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '525ff740aac8453ab3bde4fd422aebc2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 663.283060] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453258, 'name': CreateVM_Task, 'duration_secs': 0.500061} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.283235] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 663.283783] env[65385]: WARNING neutronclient.v2_0.client [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.284227] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.284439] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.284770] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 663.289705] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e84e0f66-b0da-44a1-855d-0c54b73f54b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.294989] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 663.294989] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52749277-a0d9-233c-8cfa-ecb224f505d7" [ 663.294989] env[65385]: _type = "Task" [ 663.294989] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.310205] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52749277-a0d9-233c-8cfa-ecb224f505d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.438490] env[65385]: DEBUG oslo_concurrency.lockutils [req-1d91d275-47b5-46c5-8f6e-ba82f6770e0e req-21085892-d91e-4a89-ba7a-51b9cf37ac55 service nova] Releasing lock "refresh_cache-de8632e0-bf38-440d-b6a3-895efeef122d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.542871] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Releasing lock "refresh_cache-69fdd5df-a9f2-486f-8a79-87c034366083" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.544387] env[65385]: DEBUG nova.compute.manager [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Received event network-changed-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 663.544387] env[65385]: DEBUG nova.compute.manager [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Refreshing instance network info cache due to event network-changed-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 663.544387] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Acquiring lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.544387] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Acquired lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.544551] env[65385]: DEBUG nova.network.neutron [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Refreshing network info cache for port 4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 663.617140] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb31723d-22a9-4449-9f1b-6392efaaf297 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.627372] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf80c218-c658-4b1f-9f68-261dcd739267 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.672721] env[65385]: DEBUG nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 663.676472] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec60856-2c82-47d9-a092-2496626b3cac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.686226] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e12579b-c57c-431c-80e8-390bf0b2d036 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.705164] env[65385]: DEBUG nova.compute.provider_tree [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.719895] env[65385]: DEBUG nova.network.neutron [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Successfully created port: 8227e6e0-c35a-450c-8bf3-f9b038dd9116 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 663.727876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.727876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.727876] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 663.729071] env[65385]: DEBUG oslo_vmware.api [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453254, 'name': PowerOnVM_Task, 'duration_secs': 1.060361} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.730090] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 663.730274] env[65385]: INFO nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Took 12.61 seconds to spawn the instance on the hypervisor. [ 663.730993] env[65385]: DEBUG nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 663.731308] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4683a79b-f229-4116-b8bb-d7bdb5cefff5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.754697] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453259, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.791563] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.795114] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.811278] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52749277-a0d9-233c-8cfa-ecb224f505d7, 'name': SearchDatastore_Task, 'duration_secs': 0.012275} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.811605] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.811967] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 663.812309] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.812465] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.812647] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 663.812960] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abdd337a-47f5-4dd3-9883-579e9fa26308 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.824824] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 663.824824] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 663.826338] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-248a09c3-9947-43d0-94ac-936f0269e53c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.832582] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 663.832582] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52235b65-dff4-1130-7ba3-d9fb2fef5995" [ 663.832582] env[65385]: _type = "Task" [ 663.832582] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.842809] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52235b65-dff4-1130-7ba3-d9fb2fef5995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.047704] env[65385]: WARNING neutronclient.v2_0.client [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.048444] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.048910] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.208609] env[65385]: DEBUG nova.scheduler.client.report [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 664.235747] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.236272] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.264084] env[65385]: INFO nova.compute.manager [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Took 31.99 seconds to build instance. [ 664.266916] env[65385]: DEBUG oslo_vmware.api [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453259, 'name': PowerOnVM_Task, 'duration_secs': 0.540116} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.269632] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 664.269632] env[65385]: INFO nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Took 10.03 seconds to spawn the instance on the hypervisor. [ 664.269632] env[65385]: DEBUG nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 664.269632] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691306e3-dcf3-4c3b-921e-910ee4b63a12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.295249] env[65385]: DEBUG nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 664.344070] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52235b65-dff4-1130-7ba3-d9fb2fef5995, 'name': SearchDatastore_Task, 'duration_secs': 0.013278} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.344865] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b889a31-619c-4356-b740-47e4de41d50d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.352902] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 664.352902] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b4b2a6-f520-9fda-0e7c-8550e9bdc33a" [ 664.352902] env[65385]: _type = "Task" [ 664.352902] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.362849] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b4b2a6-f520-9fda-0e7c-8550e9bdc33a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.686298] env[65385]: DEBUG nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 664.711989] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 664.715072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.080s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.719479] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 664.719807] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 664.719896] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 664.720019] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 664.720170] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 664.720312] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 664.720524] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.720673] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 664.720831] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 664.720994] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 664.721190] env[65385]: DEBUG nova.virt.hardware [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 664.721808] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.042s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.723438] env[65385]: INFO nova.compute.claims [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.726644] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ed201f-3afd-47fc-8f91-3d8d05cfc3c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.738899] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a21e4b1-d734-4939-a2ed-c15c7cc690e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.760149] env[65385]: INFO nova.scheduler.client.report [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Deleted allocations for instance c5b85574-df14-4a7a-ada0-c9668bf18a33 [ 664.768458] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a502d40a-8631-40ac-b9c5-74a574f964ab tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "69fdd5df-a9f2-486f-8a79-87c034366083" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.533s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.791978] env[65385]: INFO nova.compute.manager [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Took 29.97 seconds to build instance. [ 664.825331] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.854144] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.854591] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.873332] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b4b2a6-f520-9fda-0e7c-8550e9bdc33a, 'name': SearchDatastore_Task, 'duration_secs': 0.018312} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.873656] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.873961] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 89f3c815-8671-47ce-9e74-bf6e652bb3c3/89f3c815-8671-47ce-9e74-bf6e652bb3c3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 664.874316] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b70beef-a4a8-45f9-b3e3-99e1186d81b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.882470] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 664.882470] env[65385]: value = "task-4453260" [ 664.882470] env[65385]: _type = "Task" [ 664.882470] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.894358] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.947186] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.947745] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.048430] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.049084] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.049424] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.078726] env[65385]: WARNING neutronclient.v2_0.client [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.080334] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.080841] env[65385]: WARNING openstack [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.270337] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e704211c-ac18-47de-b353-0c3b9dffc001 tempest-ServerDiagnosticsV248Test-629729915 tempest-ServerDiagnosticsV248Test-629729915-project-member] Lock "c5b85574-df14-4a7a-ada0-c9668bf18a33" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.868s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.294655] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be55620e-8fcb-4420-a12d-d5af28011e81 tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "de8632e0-bf38-440d-b6a3-895efeef122d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.481s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.394678] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453260, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.415000] env[65385]: DEBUG nova.network.neutron [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Updated VIF entry in instance network info cache for port 4343bbc8-b6b2-4198-9f4c-d6da5ff242dc. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 665.415939] env[65385]: DEBUG nova.network.neutron [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Updating instance_info_cache with network_info: [{"id": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "address": "fa:16:3e:89:46:11", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4343bbc8-b6", "ovs_interfaceid": "4343bbc8-b6b2-4198-9f4c-d6da5ff242dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.418804] env[65385]: DEBUG nova.network.neutron [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Successfully updated port: 8227e6e0-c35a-450c-8bf3-f9b038dd9116 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 665.444213] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.444213] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.552394] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.554126] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.554900] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.688401] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.688726] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.693262] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.693625] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.770037] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.770684] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.771159] env[65385]: WARNING openstack [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.906211] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636025} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.909106] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 89f3c815-8671-47ce-9e74-bf6e652bb3c3/89f3c815-8671-47ce-9e74-bf6e652bb3c3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 665.909359] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 665.910293] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b6f6313-83b5-4e93-9c40-60212e7b01c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.918296] env[65385]: DEBUG nova.network.neutron [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updating instance_info_cache with network_info: [{"id": "b89f2175-8529-42c4-9a78-6e7116ad7293", "address": "fa:16:3e:20:e7:94", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.58", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb89f2175-85", "ovs_interfaceid": "b89f2175-8529-42c4-9a78-6e7116ad7293", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "address": "fa:16:3e:85:ae:47", "network": {"id": "6646b1ca-4991-4816-bbbf-630b1414ab64", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-120680701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a0b59-d6", "ovs_interfaceid": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c68958e-7ea9-4a06-b35a-05c48985738b", "address": "fa:16:3e:ec:18:75", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c68958e-7e", "ovs_interfaceid": "8c68958e-7ea9-4a06-b35a-05c48985738b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.921235] env[65385]: DEBUG oslo_concurrency.lockutils [req-592c5ed9-1966-4342-8faf-7df7c6312855 req-2e52cc41-5b54-413d-bf04-96d0d161d1cf service nova] Releasing lock "refresh_cache-427d5c6d-ab01-42ef-8ee5-edd608896b8b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 665.923024] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "refresh_cache-a71d2fa6-006c-4f76-8ae8-467e149297da" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.923187] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquired lock "refresh_cache-a71d2fa6-006c-4f76-8ae8-467e149297da" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.923354] env[65385]: DEBUG nova.network.neutron [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 665.927649] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 665.927649] env[65385]: value = "task-4453261" [ 665.927649] env[65385]: _type = "Task" [ 665.927649] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.942390] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.083908] env[65385]: DEBUG nova.compute.manager [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Received event network-vif-plugged-18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 666.084218] env[65385]: DEBUG oslo_concurrency.lockutils [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Acquiring lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.084685] env[65385]: DEBUG oslo_concurrency.lockutils [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.084685] env[65385]: DEBUG oslo_concurrency.lockutils [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.084827] env[65385]: DEBUG nova.compute.manager [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] No waiting events found dispatching network-vif-plugged-18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 666.085037] env[65385]: WARNING nova.compute.manager [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Received unexpected event network-vif-plugged-18c18d61-0493-40e3-8883-c90faabf147b for instance with vm_state building and task_state spawning. [ 666.085206] env[65385]: DEBUG nova.compute.manager [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Received event network-changed-18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 666.085357] env[65385]: DEBUG nova.compute.manager [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Refreshing instance network info cache due to event network-changed-18c18d61-0493-40e3-8883-c90faabf147b. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 666.085667] env[65385]: DEBUG oslo_concurrency.lockutils [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Acquiring lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.085829] env[65385]: DEBUG oslo_concurrency.lockutils [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Acquired lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.085974] env[65385]: DEBUG nova.network.neutron [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Refreshing network info cache for port 18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 666.096589] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-vif-plugged-b89f2175-8529-42c4-9a78-6e7116ad7293 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 666.096914] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquiring lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.097220] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.097531] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.097751] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] No waiting events found dispatching network-vif-plugged-b89f2175-8529-42c4-9a78-6e7116ad7293 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 666.097944] env[65385]: WARNING nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received unexpected event network-vif-plugged-b89f2175-8529-42c4-9a78-6e7116ad7293 for instance with vm_state building and task_state spawning. [ 666.098221] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-changed-b89f2175-8529-42c4-9a78-6e7116ad7293 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 666.098392] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Refreshing instance network info cache due to event network-changed-b89f2175-8529-42c4-9a78-6e7116ad7293. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 666.098704] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquiring lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.201201] env[65385]: DEBUG nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 666.211398] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870ac22f-3b42-4e36-bdec-c9d59e85696f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.218992] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102fbe0f-e4f8-42de-b528-785f349134a0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.264762] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561949de-5d87-49ec-a75f-6fa93feb4613 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.276735] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5551bf-8978-4a8f-8d49-586c5dca5d11 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.298697] env[65385]: DEBUG nova.compute.provider_tree [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.388950] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "42b9edf6-1873-49c3-8074-8eef654ac371" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.389281] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.389925] env[65385]: DEBUG nova.compute.manager [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Going to confirm migration 1 {{(pid=65385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 666.421730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Releasing lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.422194] env[65385]: DEBUG nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Instance network_info: |[{"id": "b89f2175-8529-42c4-9a78-6e7116ad7293", "address": "fa:16:3e:20:e7:94", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.58", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb89f2175-85", "ovs_interfaceid": "b89f2175-8529-42c4-9a78-6e7116ad7293", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "address": "fa:16:3e:85:ae:47", "network": {"id": "6646b1ca-4991-4816-bbbf-630b1414ab64", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-120680701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a0b59-d6", "ovs_interfaceid": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c68958e-7ea9-4a06-b35a-05c48985738b", "address": "fa:16:3e:ec:18:75", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c68958e-7e", "ovs_interfaceid": "8c68958e-7ea9-4a06-b35a-05c48985738b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 666.422829] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquired lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.423288] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Refreshing network info cache for port b89f2175-8529-42c4-9a78-6e7116ad7293 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 666.426117] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:e7:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b89f2175-8529-42c4-9a78-6e7116ad7293', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:ae:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f23a0b59-d6a0-416e-a41c-d0f1f12672ee', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:18:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c68958e-7ea9-4a06-b35a-05c48985738b', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.438092] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Creating folder: Project (21132a4099d84aaa89698915f7e91c01). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.440327] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.440676] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.448455] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c66bdbb-35fd-4da7-beda-74ef6ae1622d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.457860] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.458114] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.473703] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086828} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.475682] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 666.475995] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Created folder: Project (21132a4099d84aaa89698915f7e91c01) in parent group-v870881. [ 666.476194] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Creating folder: Instances. Parent ref: group-v870948. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.477271] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3558f3-12f5-40e8-a78f-e883b4888aee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.479923] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aac36485-b0fc-4831-a11a-9239173de3d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.508506] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] 89f3c815-8671-47ce-9e74-bf6e652bb3c3/89f3c815-8671-47ce-9e74-bf6e652bb3c3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 666.510764] env[65385]: DEBUG nova.network.neutron [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 666.515035] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c68d5223-3a42-48bf-aab5-d1db02961b5d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.528728] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Created folder: Instances in parent group-v870948. [ 666.529039] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 666.529492] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 666.530109] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f5856aa-5599-46a7-a07f-e61dde4ca326 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.551143] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.551534] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.560666] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 666.560666] env[65385]: value = "task-4453264" [ 666.560666] env[65385]: _type = "Task" [ 666.560666] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.568201] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.568201] env[65385]: value = "task-4453265" [ 666.568201] env[65385]: _type = "Task" [ 666.568201] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.576397] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453264, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.583057] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453265, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.589011] env[65385]: WARNING neutronclient.v2_0.client [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.589813] env[65385]: WARNING openstack [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.590383] env[65385]: WARNING openstack [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.703038] env[65385]: WARNING neutronclient.v2_0.client [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.703866] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.704246] env[65385]: WARNING openstack [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.733556] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.800026] env[65385]: DEBUG nova.scheduler.client.report [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 666.896600] env[65385]: WARNING neutronclient.v2_0.client [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.942236] env[65385]: WARNING neutronclient.v2_0.client [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.942236] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.942236] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.961499] env[65385]: DEBUG nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 667.075646] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453264, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.083856] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453265, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.134387] env[65385]: WARNING openstack [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.134812] env[65385]: WARNING openstack [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.201087] env[65385]: WARNING neutronclient.v2_0.client [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.201392] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.201724] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.201983] env[65385]: DEBUG nova.network.neutron [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 667.202983] env[65385]: DEBUG nova.objects.instance [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lazy-loading 'info_cache' on Instance uuid 42b9edf6-1873-49c3-8074-8eef654ac371 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 667.275213] env[65385]: DEBUG nova.network.neutron [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Updating instance_info_cache with network_info: [{"id": "8227e6e0-c35a-450c-8bf3-f9b038dd9116", "address": "fa:16:3e:76:87:00", "network": {"id": "d4a6d91f-904c-4d2d-957b-567824e8cbf4", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1383487111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "525ff740aac8453ab3bde4fd422aebc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bed837fa-6b6a-4192-a229-a99426a46065", "external-id": "nsx-vlan-transportzone-954", "segmentation_id": 954, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8227e6e0-c3", "ovs_interfaceid": "8227e6e0-c35a-450c-8bf3-f9b038dd9116", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.307186] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.309632] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.567s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.311835] env[65385]: INFO nova.compute.claims [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.388605] env[65385]: WARNING neutronclient.v2_0.client [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.390012] env[65385]: WARNING openstack [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.391152] env[65385]: WARNING openstack [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.493976] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.497648] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.498466] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.535852] env[65385]: DEBUG nova.network.neutron [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updated VIF entry in instance network info cache for port 18c18d61-0493-40e3-8883-c90faabf147b. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 667.536450] env[65385]: DEBUG nova.network.neutron [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.579191] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453264, 'name': ReconfigVM_Task, 'duration_secs': 0.897344} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.587028] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Reconfigured VM instance instance-00000019 to attach disk [datastore2] 89f3c815-8671-47ce-9e74-bf6e652bb3c3/89f3c815-8671-47ce-9e74-bf6e652bb3c3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 667.587028] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453265, 'name': CreateVM_Task, 'duration_secs': 0.872387} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.587581] env[65385]: WARNING neutronclient.v2_0.client [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.588759] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.589385] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.597358] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3cadb7b-32e1-4635-b9fb-a7a30b1b2f18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.599195] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 667.600271] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.600993] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.601407] env[65385]: WARNING neutronclient.v2_0.client [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.601613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.601810] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.602916] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 667.602916] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe6257a6-0f99-475c-b6a9-ed7349f1ba33 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.609933] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 667.609933] env[65385]: value = "task-4453266" [ 667.609933] env[65385]: _type = "Task" [ 667.609933] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.612094] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 667.612094] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fefe3-d934-9aa7-5f82-48e6ec58c0db" [ 667.612094] env[65385]: _type = "Task" [ 667.612094] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.627445] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453266, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.631033] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fefe3-d934-9aa7-5f82-48e6ec58c0db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.723429] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updated VIF entry in instance network info cache for port b89f2175-8529-42c4-9a78-6e7116ad7293. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 667.723894] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updating instance_info_cache with network_info: [{"id": "b89f2175-8529-42c4-9a78-6e7116ad7293", "address": "fa:16:3e:20:e7:94", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.58", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb89f2175-85", "ovs_interfaceid": "b89f2175-8529-42c4-9a78-6e7116ad7293", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "address": "fa:16:3e:85:ae:47", "network": {"id": "6646b1ca-4991-4816-bbbf-630b1414ab64", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-120680701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a0b59-d6", "ovs_interfaceid": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c68958e-7ea9-4a06-b35a-05c48985738b", "address": "fa:16:3e:ec:18:75", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c68958e-7e", "ovs_interfaceid": "8c68958e-7ea9-4a06-b35a-05c48985738b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.778334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Releasing lock "refresh_cache-a71d2fa6-006c-4f76-8ae8-467e149297da" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.778721] env[65385]: DEBUG nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Instance network_info: |[{"id": "8227e6e0-c35a-450c-8bf3-f9b038dd9116", "address": "fa:16:3e:76:87:00", "network": {"id": "d4a6d91f-904c-4d2d-957b-567824e8cbf4", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1383487111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "525ff740aac8453ab3bde4fd422aebc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bed837fa-6b6a-4192-a229-a99426a46065", "external-id": "nsx-vlan-transportzone-954", "segmentation_id": 954, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8227e6e0-c3", "ovs_interfaceid": "8227e6e0-c35a-450c-8bf3-f9b038dd9116", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 667.780461] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:87:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bed837fa-6b6a-4192-a229-a99426a46065', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8227e6e0-c35a-450c-8bf3-f9b038dd9116', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 667.788943] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Creating folder: Project (525ff740aac8453ab3bde4fd422aebc2). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.789180] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4bd0210-b286-431b-acfc-2714dffbb02a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.801673] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Created folder: Project (525ff740aac8453ab3bde4fd422aebc2) in parent group-v870881. [ 667.802399] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Creating folder: Instances. Parent ref: group-v870951. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.802399] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02d52f72-a901-40ba-b6ee-a66e89315bbc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.813715] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Created folder: Instances in parent group-v870951. [ 667.814066] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 667.814294] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 667.814544] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95dafae7-89ff-4706-84cc-f455a48ad2ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.832740] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "e9fa73fa-542f-4c12-b09c-b5b1f2fdea34" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.833011] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "e9fa73fa-542f-4c12-b09c-b5b1f2fdea34" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.846337] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 667.846337] env[65385]: value = "task-4453269" [ 667.846337] env[65385]: _type = "Task" [ 667.846337] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.858262] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453269, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.893715] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "28fd3828-38ec-426f-b4dc-1e094304780b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.894019] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.048408] env[65385]: DEBUG oslo_concurrency.lockutils [req-2354e592-ef19-4fb0-802f-417ece9220de req-9a5c0a13-b063-48bc-895a-2eba7ee5bd41 service nova] Releasing lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.128569] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453266, 'name': Rename_Task, 'duration_secs': 0.332408} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.134024] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 668.134024] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fefe3-d934-9aa7-5f82-48e6ec58c0db, 'name': SearchDatastore_Task, 'duration_secs': 0.033486} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.134024] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ac29df8-8e0a-4fa6-971e-aa657fa44c2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.135425] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.135751] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 668.136046] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.136239] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.138034] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 668.138034] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74a72f67-7916-480f-972c-947206385183 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.145933] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 668.145933] env[65385]: value = "task-4453270" [ 668.145933] env[65385]: _type = "Task" [ 668.145933] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.151217] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 668.151217] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 668.152360] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96036ad2-d3f0-4c5c-a3cd-fe4cef0efd31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.158637] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453270, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.163020] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 668.163020] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528da457-d404-7569-e98d-99fad5e2e904" [ 668.163020] env[65385]: _type = "Task" [ 668.163020] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.175476] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528da457-d404-7569-e98d-99fad5e2e904, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.210991] env[65385]: WARNING neutronclient.v2_0.client [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.211725] env[65385]: WARNING openstack [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.212626] env[65385]: WARNING openstack [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.229073] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Releasing lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.229271] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-vif-plugged-f23a0b59-d6a0-416e-a41c-d0f1f12672ee {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 668.229661] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquiring lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.229661] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.230525] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.230525] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] No waiting events found dispatching network-vif-plugged-f23a0b59-d6a0-416e-a41c-d0f1f12672ee {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 668.230525] env[65385]: WARNING nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received unexpected event network-vif-plugged-f23a0b59-d6a0-416e-a41c-d0f1f12672ee for instance with vm_state building and task_state spawning. [ 668.231105] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-changed-f23a0b59-d6a0-416e-a41c-d0f1f12672ee {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 668.231609] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Refreshing instance network info cache due to event network-changed-f23a0b59-d6a0-416e-a41c-d0f1f12672ee. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 668.231609] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquiring lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.231745] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquired lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.231804] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Refreshing network info cache for port f23a0b59-d6a0-416e-a41c-d0f1f12672ee {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 668.340470] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "e9fa73fa-542f-4c12-b09c-b5b1f2fdea34" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.507s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.341283] env[65385]: DEBUG nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 668.358919] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453269, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.367655] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "a8c76329-5056-4cf0-ac32-85d46429d3db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.367916] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.368179] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "a8c76329-5056-4cf0-ac32-85d46429d3db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.368367] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.368533] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.370976] env[65385]: INFO nova.compute.manager [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Terminating instance [ 668.401788] env[65385]: DEBUG nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 668.561987] env[65385]: WARNING openstack [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.563057] env[65385]: WARNING openstack [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.659537] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453270, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.660923] env[65385]: WARNING neutronclient.v2_0.client [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.661554] env[65385]: WARNING openstack [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.661900] env[65385]: WARNING openstack [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.684785] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528da457-d404-7569-e98d-99fad5e2e904, 'name': SearchDatastore_Task, 'duration_secs': 0.013132} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.685960] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-809a6845-881f-4234-b26a-251fa599d0d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.696664] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 668.696664] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525c0190-acf9-948d-5df2-126a4b18b252" [ 668.696664] env[65385]: _type = "Task" [ 668.696664] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.706365] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525c0190-acf9-948d-5df2-126a4b18b252, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.737443] env[65385]: WARNING neutronclient.v2_0.client [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.738409] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.738723] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.785278] env[65385]: DEBUG nova.network.neutron [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance_info_cache with network_info: [{"id": "03771ba6-a616-4715-afb9-19306b1b0903", "address": "fa:16:3e:2b:4a:04", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03771ba6-a6", "ovs_interfaceid": "03771ba6-a616-4715-afb9-19306b1b0903", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 668.814540] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73e11ec-aeca-43db-98cd-cca037f981d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.825547] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5bbf90-b7ce-4632-93bb-7032a3af6257 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.858941] env[65385]: DEBUG nova.compute.utils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 668.861819] env[65385]: DEBUG nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 668.862025] env[65385]: DEBUG nova.network.neutron [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 668.862484] env[65385]: WARNING neutronclient.v2_0.client [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.863132] env[65385]: WARNING neutronclient.v2_0.client [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.863793] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.864150] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.873946] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f39443-4a55-4bdd-83e7-f56fefc44608 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.879792] env[65385]: DEBUG nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 668.885540] env[65385]: DEBUG nova.compute.manager [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 668.885540] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 668.885540] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0462b39-cc2e-4e9f-b93a-2b129a2b8243 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.892620] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453269, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.902570] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1a924c-002b-485b-862d-a1823597e216 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.913537] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 668.917798] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e7d4635-67eb-4b45-bd5c-9d518f24e406 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.929651] env[65385]: DEBUG nova.compute.provider_tree [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.932809] env[65385]: DEBUG oslo_vmware.api [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 668.932809] env[65385]: value = "task-4453271" [ 668.932809] env[65385]: _type = "Task" [ 668.932809] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.937476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.944399] env[65385]: DEBUG oslo_vmware.api [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.966313] env[65385]: DEBUG nova.policy [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a484b18f526c4555a71600c76a4ca6b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec2768a4aefb4e1b9484fb266d45d2a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 668.998996] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.999407] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.159697] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453270, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.210984] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525c0190-acf9-948d-5df2-126a4b18b252, 'name': SearchDatastore_Task, 'duration_secs': 0.03329} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.211198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.211533] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5291c04c-24ad-4c64-9fc2-fde42da4bc0f/5291c04c-24ad-4c64-9fc2-fde42da4bc0f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 669.211799] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f07d4a5b-ce0e-4463-9103-67a54fe36107 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.222669] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 669.222669] env[65385]: value = "task-4453272" [ 669.222669] env[65385]: _type = "Task" [ 669.222669] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.231053] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453272, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.290682] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-42b9edf6-1873-49c3-8074-8eef654ac371" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.291212] env[65385]: DEBUG nova.objects.instance [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lazy-loading 'migration_context' on Instance uuid 42b9edf6-1873-49c3-8074-8eef654ac371 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 669.379029] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453269, 'name': CreateVM_Task, 'duration_secs': 1.530042} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.379029] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 669.379029] env[65385]: WARNING neutronclient.v2_0.client [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.379029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.379029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.379545] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 669.379715] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a613dd4-0d5c-4555-afb0-608f01977934 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.384592] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.384874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.385444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.385444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.385726] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.391787] env[65385]: INFO nova.compute.manager [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Terminating instance [ 669.397649] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 669.397649] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5297643f-27a7-8fa3-e28a-294cce77a66f" [ 669.397649] env[65385]: _type = "Task" [ 669.397649] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.405366] env[65385]: DEBUG nova.network.neutron [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Successfully created port: 07665fa0-2bf4-4c1d-9058-505e323caa56 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 669.417075] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5297643f-27a7-8fa3-e28a-294cce77a66f, 'name': SearchDatastore_Task, 'duration_secs': 0.011819} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.418107] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.418107] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 669.418491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.420683] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.420683] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 669.420683] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb60cb45-2904-4cab-91b7-79575f8cee54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.437036] env[65385]: DEBUG nova.scheduler.client.report [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 669.441419] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 669.441821] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 669.459657] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3d03dda-80e3-4398-8e79-564d07be7fec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.471172] env[65385]: DEBUG oslo_vmware.api [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453271, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.472591] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 669.472591] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52468b37-38b2-9f85-0f81-594e5925c345" [ 669.472591] env[65385]: _type = "Task" [ 669.472591] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.486991] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52468b37-38b2-9f85-0f81-594e5925c345, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.663247] env[65385]: DEBUG oslo_vmware.api [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453270, 'name': PowerOnVM_Task, 'duration_secs': 1.365955} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.663589] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 669.663955] env[65385]: INFO nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Took 10.73 seconds to spawn the instance on the hypervisor. [ 669.663955] env[65385]: DEBUG nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 669.665325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78455143-8605-4bc4-98b0-8eb796066ba7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.734201] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453272, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.781199] env[65385]: DEBUG nova.compute.manager [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Received event network-vif-plugged-8227e6e0-c35a-450c-8bf3-f9b038dd9116 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 669.781199] env[65385]: DEBUG oslo_concurrency.lockutils [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Acquiring lock "a71d2fa6-006c-4f76-8ae8-467e149297da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.781199] env[65385]: DEBUG oslo_concurrency.lockutils [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.781199] env[65385]: DEBUG oslo_concurrency.lockutils [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.781199] env[65385]: DEBUG nova.compute.manager [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] No waiting events found dispatching network-vif-plugged-8227e6e0-c35a-450c-8bf3-f9b038dd9116 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 669.781986] env[65385]: WARNING nova.compute.manager [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Received unexpected event network-vif-plugged-8227e6e0-c35a-450c-8bf3-f9b038dd9116 for instance with vm_state building and task_state spawning. [ 669.781986] env[65385]: DEBUG nova.compute.manager [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Received event network-changed-8227e6e0-c35a-450c-8bf3-f9b038dd9116 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 669.781986] env[65385]: DEBUG nova.compute.manager [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Refreshing instance network info cache due to event network-changed-8227e6e0-c35a-450c-8bf3-f9b038dd9116. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 669.781986] env[65385]: DEBUG oslo_concurrency.lockutils [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Acquiring lock "refresh_cache-a71d2fa6-006c-4f76-8ae8-467e149297da" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.782208] env[65385]: DEBUG oslo_concurrency.lockutils [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Acquired lock "refresh_cache-a71d2fa6-006c-4f76-8ae8-467e149297da" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.782208] env[65385]: DEBUG nova.network.neutron [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Refreshing network info cache for port 8227e6e0-c35a-450c-8bf3-f9b038dd9116 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 669.795482] env[65385]: DEBUG nova.objects.base [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Object Instance<42b9edf6-1873-49c3-8074-8eef654ac371> lazy-loaded attributes: info_cache,migration_context {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 669.796802] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b128b5-f18d-46af-98cb-2dec015fec17 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.819723] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1167daf-b8d7-44ac-905b-841ca7e7cf22 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.826802] env[65385]: DEBUG oslo_vmware.api [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 669.826802] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526f8aba-e49a-fe2a-85cd-3c245cc97404" [ 669.826802] env[65385]: _type = "Task" [ 669.826802] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.837676] env[65385]: DEBUG oslo_vmware.api [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526f8aba-e49a-fe2a-85cd-3c245cc97404, 'name': SearchDatastore_Task, 'duration_secs': 0.008838} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.838102] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.898190] env[65385]: DEBUG nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 669.900961] env[65385]: DEBUG nova.compute.manager [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 669.901284] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.902478] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9887c3d-cbe0-4b4f-b6a1-41ea3bf65049 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.911738] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 669.912022] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c4ae96b-9df6-4d66-b713-e884947241cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.922999] env[65385]: DEBUG oslo_vmware.api [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 669.922999] env[65385]: value = "task-4453273" [ 669.922999] env[65385]: _type = "Task" [ 669.922999] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.925537] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 669.925854] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 669.926101] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 669.926310] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 669.926912] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 669.926912] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 669.926912] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.926912] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 669.927434] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 669.927434] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 669.927434] env[65385]: DEBUG nova.virt.hardware [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 669.928323] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ca91fd-41c1-4d12-b0db-45d64db4da0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.951954] env[65385]: DEBUG oslo_vmware.api [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453273, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.952938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.953720] env[65385]: DEBUG nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 669.961032] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.246s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.961032] env[65385]: DEBUG nova.objects.instance [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lazy-loading 'resources' on Instance uuid 065659ae-2163-4d6b-a905-fc99d12b6790 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 669.962307] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab089454-d1e7-4070-b604-074139ee5b76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.975660] env[65385]: DEBUG oslo_vmware.api [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453271, 'name': PowerOffVM_Task, 'duration_secs': 0.773041} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.980299] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 669.980545] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 669.981194] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d63a3029-faf3-4d96-9e04-7a583137573f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.003144] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52468b37-38b2-9f85-0f81-594e5925c345, 'name': SearchDatastore_Task, 'duration_secs': 0.048379} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.004092] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd53c32c-cc1d-4223-a032-55f9fc23d35a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.012113] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 670.012113] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5213dd6b-56d8-b3d3-86a5-c2b2b91911b3" [ 670.012113] env[65385]: _type = "Task" [ 670.012113] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.030709] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5213dd6b-56d8-b3d3-86a5-c2b2b91911b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.086038] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 670.086185] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 670.086463] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Deleting the datastore file [datastore1] a8c76329-5056-4cf0-ac32-85d46429d3db {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 670.086780] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e03815c1-de8f-4ffb-941a-491891b0228d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.097792] env[65385]: DEBUG oslo_vmware.api [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for the task: (returnval){ [ 670.097792] env[65385]: value = "task-4453275" [ 670.097792] env[65385]: _type = "Task" [ 670.097792] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.118618] env[65385]: DEBUG oslo_vmware.api [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.207258] env[65385]: INFO nova.compute.manager [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Took 29.11 seconds to build instance. [ 670.242280] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453272, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55595} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.243241] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5291c04c-24ad-4c64-9fc2-fde42da4bc0f/5291c04c-24ad-4c64-9fc2-fde42da4bc0f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 670.243241] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 670.243241] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-786f5f83-aeab-4db3-baea-0ab770df82d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.251744] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 670.251744] env[65385]: value = "task-4453276" [ 670.251744] env[65385]: _type = "Task" [ 670.251744] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.270671] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453276, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.285182] env[65385]: WARNING neutronclient.v2_0.client [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.286476] env[65385]: WARNING openstack [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.286476] env[65385]: WARNING openstack [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.331321] env[65385]: WARNING neutronclient.v2_0.client [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.332033] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.332359] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.439737] env[65385]: DEBUG oslo_vmware.api [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453273, 'name': PowerOffVM_Task, 'duration_secs': 0.245367} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.439973] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 670.440230] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 670.440517] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9405701-fc55-4ae0-b21f-81aeb5c6db9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.476404] env[65385]: DEBUG nova.compute.utils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 670.477906] env[65385]: DEBUG nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 670.482045] env[65385]: DEBUG nova.network.neutron [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 670.482323] env[65385]: WARNING neutronclient.v2_0.client [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.482669] env[65385]: WARNING neutronclient.v2_0.client [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.488030] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.488030] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.532334] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5213dd6b-56d8-b3d3-86a5-c2b2b91911b3, 'name': SearchDatastore_Task, 'duration_secs': 0.015476} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.534047] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.534339] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] a71d2fa6-006c-4f76-8ae8-467e149297da/a71d2fa6-006c-4f76-8ae8-467e149297da.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 670.534548] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 670.534799] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 670.534896] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Deleting the datastore file [datastore1] bf79fbcd-e943-4ff3-bcf8-6a2817df8292 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 670.536217] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e953261f-10e5-44ac-811f-ae7735d5b53f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.537429] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b024f2e-0f4c-47e4-bced-47319e31785a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.550885] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 670.550885] env[65385]: value = "task-4453279" [ 670.550885] env[65385]: _type = "Task" [ 670.550885] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.550885] env[65385]: DEBUG oslo_vmware.api [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for the task: (returnval){ [ 670.550885] env[65385]: value = "task-4453278" [ 670.550885] env[65385]: _type = "Task" [ 670.550885] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.573675] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453279, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.573954] env[65385]: DEBUG oslo_vmware.api [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453278, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.615673] env[65385]: DEBUG oslo_vmware.api [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Task: {'id': task-4453275, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254736} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.616144] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 670.616586] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 670.616586] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.616586] env[65385]: INFO nova.compute.manager [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Took 1.73 seconds to destroy the instance on the hypervisor. [ 670.617050] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 670.617212] env[65385]: DEBUG nova.compute.manager [-] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 670.618672] env[65385]: DEBUG nova.network.neutron [-] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 670.618672] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.618672] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.618672] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.678154] env[65385]: DEBUG nova.policy [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16d2d13d8f4d4df0b0507fb6efd0faa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8167324e3664a578a5ca8c8a04c1241', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 670.713718] env[65385]: DEBUG oslo_concurrency.lockutils [None req-73836df9-d100-4641-96f9-1f949ab084c8 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.626s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.762639] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453276, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080054} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.765677] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 670.766628] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793f93ab-aa44-43a7-9464-7fadc0b64b7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.794994] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] 5291c04c-24ad-4c64-9fc2-fde42da4bc0f/5291c04c-24ad-4c64-9fc2-fde42da4bc0f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 670.798814] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e584256-fb37-4378-94ed-a248f1ee0e0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.816699] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updated VIF entry in instance network info cache for port f23a0b59-d6a0-416e-a41c-d0f1f12672ee. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 670.817081] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updating instance_info_cache with network_info: [{"id": "b89f2175-8529-42c4-9a78-6e7116ad7293", "address": "fa:16:3e:20:e7:94", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.58", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb89f2175-85", "ovs_interfaceid": "b89f2175-8529-42c4-9a78-6e7116ad7293", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "address": "fa:16:3e:85:ae:47", "network": {"id": "6646b1ca-4991-4816-bbbf-630b1414ab64", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-120680701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a0b59-d6", "ovs_interfaceid": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c68958e-7ea9-4a06-b35a-05c48985738b", "address": "fa:16:3e:ec:18:75", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c68958e-7e", "ovs_interfaceid": "8c68958e-7ea9-4a06-b35a-05c48985738b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 670.824496] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 670.824496] env[65385]: value = "task-4453280" [ 670.824496] env[65385]: _type = "Task" [ 670.824496] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.838020] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453280, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.842359] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.982105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c292e317-8ff2-427f-a724-22f39c857046 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.991282] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530fad98-8cfa-4dfa-ab89-d36ef0389459 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.999257] env[65385]: DEBUG nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 671.032300] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647ce112-31f8-4df8-a71f-e852653075ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.046289] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8ada15-5398-4c58-b6d8-d25f99dc5f7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.067849] env[65385]: DEBUG nova.compute.provider_tree [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 671.070633] env[65385]: WARNING openstack [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.071018] env[65385]: WARNING openstack [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.082829] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453279, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.090118] env[65385]: DEBUG oslo_vmware.api [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453278, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.110971] env[65385]: ERROR nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [req-d46fbaeb-7f0d-49a4-b656-2d3c099d7ede] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d46fbaeb-7f0d-49a4-b656-2d3c099d7ede"}]} [ 671.136342] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 671.153468] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 671.153694] env[65385]: DEBUG nova.compute.provider_tree [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 671.170393] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 671.191991] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 671.320221] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Releasing lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.320564] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-vif-plugged-8c68958e-7ea9-4a06-b35a-05c48985738b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 671.320862] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquiring lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.321179] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.321429] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.321843] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] No waiting events found dispatching network-vif-plugged-8c68958e-7ea9-4a06-b35a-05c48985738b {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 671.322155] env[65385]: WARNING nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received unexpected event network-vif-plugged-8c68958e-7ea9-4a06-b35a-05c48985738b for instance with vm_state building and task_state spawning. [ 671.322155] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-changed-8c68958e-7ea9-4a06-b35a-05c48985738b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 671.322294] env[65385]: DEBUG nova.compute.manager [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Refreshing instance network info cache due to event network-changed-8c68958e-7ea9-4a06-b35a-05c48985738b. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 671.322655] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquiring lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.322655] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Acquired lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.322894] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Refreshing network info cache for port 8c68958e-7ea9-4a06-b35a-05c48985738b {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 671.337741] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453280, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.582185] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453279, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.582453] env[65385]: DEBUG oslo_vmware.api [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Task: {'id': task-4453278, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.617457} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.582631] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 671.585932] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 671.585932] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.585932] env[65385]: INFO nova.compute.manager [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Took 1.68 seconds to destroy the instance on the hypervisor. [ 671.585932] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 671.585932] env[65385]: DEBUG nova.compute.manager [-] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 671.585932] env[65385]: DEBUG nova.network.neutron [-] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 671.585932] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.585932] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.585932] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.675139] env[65385]: DEBUG nova.network.neutron [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Successfully updated port: 07665fa0-2bf4-4c1d-9058-505e323caa56 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 671.694653] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979a0538-9c71-4539-8c31-182fe4e915b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.708126] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9037ba-848f-4880-8bb2-dd652f3e743e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.748562] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a471d880-65d1-4670-8877-53e4caf437df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.757904] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46546398-5b9b-4c7c-afdc-0cf82237f34e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.776275] env[65385]: DEBUG nova.compute.provider_tree [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 671.830847] env[65385]: WARNING neutronclient.v2_0.client [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.832038] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.835127] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.855098] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453280, 'name': ReconfigVM_Task, 'duration_secs': 0.787015} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.855433] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Reconfigured VM instance instance-00000018 to attach disk [datastore2] 5291c04c-24ad-4c64-9fc2-fde42da4bc0f/5291c04c-24ad-4c64-9fc2-fde42da4bc0f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 671.857036] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9299131-a4f1-4fc0-a313-1fab41f03283 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.864568] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 671.864568] env[65385]: value = "task-4453281" [ 671.864568] env[65385]: _type = "Task" [ 671.864568] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.875830] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453281, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.022876] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 672.041321] env[65385]: DEBUG nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 672.050531] env[65385]: DEBUG nova.network.neutron [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Successfully created port: d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 672.069523] env[65385]: WARNING neutronclient.v2_0.client [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 672.070206] env[65385]: WARNING openstack [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.071033] env[65385]: WARNING openstack [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.081594] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453279, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.051812} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.086906] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 672.087500] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 672.087853] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 672.088241] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 672.089079] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 672.089079] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 672.089079] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.089534] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 672.089866] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 672.090211] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 672.090594] env[65385]: DEBUG nova.virt.hardware [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 672.091134] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] a71d2fa6-006c-4f76-8ae8-467e149297da/a71d2fa6-006c-4f76-8ae8-467e149297da.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 672.092024] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 672.093121] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5f82eb-0d79-41fb-a5af-527f1423ba10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.096587] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb19575a-d780-4122-a595-1250112065f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.108571] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7acad85-f797-4935-beca-fea85efc5ba3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.113948] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 672.113948] env[65385]: value = "task-4453282" [ 672.113948] env[65385]: _type = "Task" [ 672.113948] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.136279] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453282, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.185288] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "refresh_cache-950efcc6-7e4f-4b30-b0fa-d940f893e1d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.185288] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquired lock "refresh_cache-950efcc6-7e4f-4b30-b0fa-d940f893e1d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.185288] env[65385]: DEBUG nova.network.neutron [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 672.199294] env[65385]: DEBUG nova.network.neutron [-] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 672.302319] env[65385]: ERROR nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] [req-f7085497-4c62-4881-9f7e-c77ee71ccbf4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f7085497-4c62-4881-9f7e-c77ee71ccbf4"}]} [ 672.330041] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 672.354782] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 672.355299] env[65385]: DEBUG nova.compute.provider_tree [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 672.374516] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 672.381234] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453281, 'name': Rename_Task, 'duration_secs': 0.322909} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.381234] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 672.381442] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30cf235b-df2c-4370-bbfc-4c1344145370 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.393262] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 672.393262] env[65385]: value = "task-4453283" [ 672.393262] env[65385]: _type = "Task" [ 672.393262] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.403057] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 672.409943] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.505886] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "3657ec21-3ec7-44c9-92d1-570655b58f26" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.506693] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.626049] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453282, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082506} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.628962] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 672.630283] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5caf3fd1-1de0-4914-a51f-4d7af3271c58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.659931] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] a71d2fa6-006c-4f76-8ae8-467e149297da/a71d2fa6-006c-4f76-8ae8-467e149297da.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 672.663515] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82bf4b88-fa46-469f-8e6b-90d383611f86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.688596] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.689724] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.697473] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 672.697473] env[65385]: value = "task-4453284" [ 672.697473] env[65385]: _type = "Task" [ 672.697473] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.701624] env[65385]: INFO nova.compute.manager [-] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Took 2.08 seconds to deallocate network for instance. [ 672.715155] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.802069] env[65385]: DEBUG nova.network.neutron [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Updated VIF entry in instance network info cache for port 8227e6e0-c35a-450c-8bf3-f9b038dd9116. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 672.802487] env[65385]: DEBUG nova.network.neutron [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Updating instance_info_cache with network_info: [{"id": "8227e6e0-c35a-450c-8bf3-f9b038dd9116", "address": "fa:16:3e:76:87:00", "network": {"id": "d4a6d91f-904c-4d2d-957b-567824e8cbf4", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1383487111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "525ff740aac8453ab3bde4fd422aebc2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bed837fa-6b6a-4192-a229-a99426a46065", "external-id": "nsx-vlan-transportzone-954", "segmentation_id": 954, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8227e6e0-c3", "ovs_interfaceid": "8227e6e0-c35a-450c-8bf3-f9b038dd9116", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 672.818788] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.819447] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.909707] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453283, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.915693] env[65385]: DEBUG nova.network.neutron [-] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 672.944406] env[65385]: DEBUG nova.network.neutron [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 672.980805] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6fc54b-edf4-4811-b3bf-34c889407f58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.994664] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10e8dbc-7948-47d8-adbc-c0d00ada403f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.030863] env[65385]: DEBUG nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 673.038283] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.038283] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.046077] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8286fc47-f80d-466c-8282-ceeca9caa442 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.055800] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486306cd-879d-4de3-8af9-a9b36b0e813d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.071895] env[65385]: DEBUG nova.compute.provider_tree [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 673.074096] env[65385]: WARNING neutronclient.v2_0.client [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.074243] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.074638] env[65385]: WARNING openstack [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.212875] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453284, 'name': ReconfigVM_Task, 'duration_secs': 0.45514} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.213185] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Reconfigured VM instance instance-0000001a to attach disk [datastore1] a71d2fa6-006c-4f76-8ae8-467e149297da/a71d2fa6-006c-4f76-8ae8-467e149297da.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 673.216287] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64a30c51-1a7e-4335-a7e9-61c4bba26801 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.216870] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 673.227207] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 673.227207] env[65385]: value = "task-4453285" [ 673.227207] env[65385]: _type = "Task" [ 673.227207] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.238124] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453285, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.306611] env[65385]: DEBUG oslo_concurrency.lockutils [req-7353e195-c6fa-446d-a178-67d7da059bba req-76e7a1d5-a8a5-4172-a5b7-a9e8228ac880 service nova] Releasing lock "refresh_cache-a71d2fa6-006c-4f76-8ae8-467e149297da" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.315798] env[65385]: WARNING neutronclient.v2_0.client [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.316861] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.317386] env[65385]: WARNING openstack [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.385309] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updated VIF entry in instance network info cache for port 8c68958e-7ea9-4a06-b35a-05c48985738b. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 673.385783] env[65385]: DEBUG nova.network.neutron [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updating instance_info_cache with network_info: [{"id": "b89f2175-8529-42c4-9a78-6e7116ad7293", "address": "fa:16:3e:20:e7:94", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.58", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb89f2175-85", "ovs_interfaceid": "b89f2175-8529-42c4-9a78-6e7116ad7293", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "address": "fa:16:3e:85:ae:47", "network": {"id": "6646b1ca-4991-4816-bbbf-630b1414ab64", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-120680701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a0b59-d6", "ovs_interfaceid": "f23a0b59-d6a0-416e-a41c-d0f1f12672ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8c68958e-7ea9-4a06-b35a-05c48985738b", "address": "fa:16:3e:ec:18:75", "network": {"id": "d871741b-f7c7-4b05-8e47-ef5f564878a7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-247287391", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c68958e-7e", "ovs_interfaceid": "8c68958e-7ea9-4a06-b35a-05c48985738b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.404114] env[65385]: DEBUG oslo_vmware.api [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453283, 'name': PowerOnVM_Task, 'duration_secs': 0.549103} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.404385] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 673.404608] env[65385]: INFO nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Took 16.59 seconds to spawn the instance on the hypervisor. [ 673.404780] env[65385]: DEBUG nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 673.405611] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e197abd3-0351-4a44-b1d2-20be6413fdad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.419306] env[65385]: INFO nova.compute.manager [-] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Took 1.84 seconds to deallocate network for instance. [ 673.513410] env[65385]: DEBUG nova.network.neutron [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Updating instance_info_cache with network_info: [{"id": "07665fa0-2bf4-4c1d-9058-505e323caa56", "address": "fa:16:3e:cd:c9:0c", "network": {"id": "1b05f7a0-d3eb-463c-bdfa-e1462758dce6", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1282242937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec2768a4aefb4e1b9484fb266d45d2a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07665fa0-2b", "ovs_interfaceid": "07665fa0-2bf4-4c1d-9058-505e323caa56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.565409] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 673.629659] env[65385]: DEBUG nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 42 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 673.629938] env[65385]: DEBUG nova.compute.provider_tree [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 42 to 43 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 673.630123] env[65385]: DEBUG nova.compute.provider_tree [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 673.738333] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453285, 'name': Rename_Task, 'duration_secs': 0.280151} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.738673] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 673.738938] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a1c94a4-744c-4506-a2a6-d5309fa75efe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.747988] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 673.747988] env[65385]: value = "task-4453286" [ 673.747988] env[65385]: _type = "Task" [ 673.747988] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.757290] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453286, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.849481] env[65385]: DEBUG nova.network.neutron [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Successfully updated port: d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 673.889564] env[65385]: DEBUG oslo_concurrency.lockutils [req-c82f6466-80ec-4d7e-9066-013e18636607 req-c0d22ccb-5f06-4886-b27b-0cb47f8639ce service nova] Releasing lock "refresh_cache-5291c04c-24ad-4c64-9fc2-fde42da4bc0f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.926859] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 673.928455] env[65385]: INFO nova.compute.manager [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Took 35.77 seconds to build instance. [ 674.016254] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Releasing lock "refresh_cache-950efcc6-7e4f-4b30-b0fa-d940f893e1d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 674.016618] env[65385]: DEBUG nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Instance network_info: |[{"id": "07665fa0-2bf4-4c1d-9058-505e323caa56", "address": "fa:16:3e:cd:c9:0c", "network": {"id": "1b05f7a0-d3eb-463c-bdfa-e1462758dce6", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1282242937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec2768a4aefb4e1b9484fb266d45d2a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07665fa0-2b", "ovs_interfaceid": "07665fa0-2bf4-4c1d-9058-505e323caa56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 674.017264] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:c9:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07665fa0-2bf4-4c1d-9058-505e323caa56', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 674.027512] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Creating folder: Project (ec2768a4aefb4e1b9484fb266d45d2a0). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 674.027902] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c65e438a-b450-40c6-b54f-1506d86fce75 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.043042] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Created folder: Project (ec2768a4aefb4e1b9484fb266d45d2a0) in parent group-v870881. [ 674.043042] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Creating folder: Instances. Parent ref: group-v870954. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 674.043272] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-201c5b0c-e51b-4c59-bb18-3feb385022d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.057570] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Created folder: Instances in parent group-v870954. [ 674.057911] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 674.058204] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 674.058552] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27913667-7ce5-4a8e-b872-468941ecc724 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.084030] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 674.084030] env[65385]: value = "task-4453289" [ 674.084030] env[65385]: _type = "Task" [ 674.084030] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.096810] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453289, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.137258] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.179s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.140373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 22.875s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.140373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.142160] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 674.142160] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.175s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.144351] env[65385]: INFO nova.compute.claims [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.149636] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d0c4ff-003a-42ee-a666-0af6541f1e04 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.160877] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff17076-d7d3-46dc-91aa-e41bb9e8ab35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.169941] env[65385]: INFO nova.scheduler.client.report [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Deleted allocations for instance 065659ae-2163-4d6b-a905-fc99d12b6790 [ 674.185087] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3f3e73-9e7e-4a95-ac06-e04d56c26710 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.201968] env[65385]: DEBUG nova.compute.manager [req-4e5e1fe0-3b20-4df8-8b80-7ce7a26e78b7 req-6d6f0d51-7644-449b-a275-bb9b33df756f service nova] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Received event network-vif-deleted-6b5cfde5-8f86-4000-90b9-5c6140ea1606 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 674.203492] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66dd8ae9-8d13-4bd8-80f8-3071308bfadf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.242030] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179256MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 674.242276] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.261618] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453286, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.315515] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Acquiring lock "8ac44124-d8cf-421c-9ae4-c943df94550a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.315515] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.332277] env[65385]: DEBUG nova.compute.manager [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Received event network-vif-plugged-07665fa0-2bf4-4c1d-9058-505e323caa56 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 674.332719] env[65385]: DEBUG oslo_concurrency.lockutils [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Acquiring lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.332961] env[65385]: DEBUG oslo_concurrency.lockutils [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.333155] env[65385]: DEBUG oslo_concurrency.lockutils [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.333324] env[65385]: DEBUG nova.compute.manager [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] No waiting events found dispatching network-vif-plugged-07665fa0-2bf4-4c1d-9058-505e323caa56 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 674.333485] env[65385]: WARNING nova.compute.manager [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Received unexpected event network-vif-plugged-07665fa0-2bf4-4c1d-9058-505e323caa56 for instance with vm_state building and task_state spawning. [ 674.333643] env[65385]: DEBUG nova.compute.manager [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Received event network-changed-07665fa0-2bf4-4c1d-9058-505e323caa56 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 674.333860] env[65385]: DEBUG nova.compute.manager [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Refreshing instance network info cache due to event network-changed-07665fa0-2bf4-4c1d-9058-505e323caa56. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 674.333970] env[65385]: DEBUG oslo_concurrency.lockutils [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Acquiring lock "refresh_cache-950efcc6-7e4f-4b30-b0fa-d940f893e1d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.334215] env[65385]: DEBUG oslo_concurrency.lockutils [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Acquired lock "refresh_cache-950efcc6-7e4f-4b30-b0fa-d940f893e1d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.334400] env[65385]: DEBUG nova.network.neutron [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Refreshing network info cache for port 07665fa0-2bf4-4c1d-9058-505e323caa56 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 674.343781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "de8632e0-bf38-440d-b6a3-895efeef122d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.344012] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "de8632e0-bf38-440d-b6a3-895efeef122d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.344297] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "de8632e0-bf38-440d-b6a3-895efeef122d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.344577] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "de8632e0-bf38-440d-b6a3-895efeef122d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.344840] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "de8632e0-bf38-440d-b6a3-895efeef122d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.347791] env[65385]: INFO nova.compute.manager [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Terminating instance [ 674.352878] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.353145] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquired lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.353418] env[65385]: DEBUG nova.network.neutron [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 674.429205] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7db13fb-7acf-404a-9724-4db49c0c8fd5 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.289s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.597251] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453289, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.688283] env[65385]: DEBUG oslo_concurrency.lockutils [None req-db9c508a-c64b-4141-aae8-f62180aed3c3 tempest-ServersAdmin275Test-1141133892 tempest-ServersAdmin275Test-1141133892-project-member] Lock "065659ae-2163-4d6b-a905-fc99d12b6790" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.458s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.693902] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "69fdd5df-a9f2-486f-8a79-87c034366083" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.694159] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "69fdd5df-a9f2-486f-8a79-87c034366083" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.694348] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "69fdd5df-a9f2-486f-8a79-87c034366083-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.694600] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "69fdd5df-a9f2-486f-8a79-87c034366083-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.694777] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "69fdd5df-a9f2-486f-8a79-87c034366083-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.696631] env[65385]: INFO nova.compute.manager [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Terminating instance [ 674.759764] env[65385]: DEBUG oslo_vmware.api [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453286, 'name': PowerOnVM_Task, 'duration_secs': 0.670295} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.759764] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 674.759764] env[65385]: INFO nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Took 10.07 seconds to spawn the instance on the hypervisor. [ 674.760033] env[65385]: DEBUG nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 674.761039] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebc80a3-5ce7-46ef-83a2-99a07ba2b015 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.820749] env[65385]: DEBUG nova.compute.utils [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 674.838234] env[65385]: WARNING neutronclient.v2_0.client [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.838987] env[65385]: WARNING openstack [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.840198] env[65385]: WARNING openstack [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.853716] env[65385]: DEBUG nova.compute.manager [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 674.854612] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 674.855176] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdaa18a6-dcc6-40df-a391-9e466dab5057 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.859730] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.859730] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.878937] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 674.879238] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e250137-9d79-4eb1-8f38-292949a4230a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.887268] env[65385]: DEBUG oslo_vmware.api [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 674.887268] env[65385]: value = "task-4453290" [ 674.887268] env[65385]: _type = "Task" [ 674.887268] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.905987] env[65385]: DEBUG oslo_vmware.api [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453290, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.931083] env[65385]: DEBUG nova.network.neutron [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 674.986124] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.986557] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.096804] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453289, 'name': CreateVM_Task, 'duration_secs': 0.525792} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.097044] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 675.097511] env[65385]: WARNING neutronclient.v2_0.client [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.097873] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.098032] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.098346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 675.098770] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31b3cf22-5e72-4c8a-a218-de9b6b127539 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.104498] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 675.104498] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bb12e-cb26-e965-241e-9b1ab829d419" [ 675.104498] env[65385]: _type = "Task" [ 675.104498] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.114178] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bb12e-cb26-e965-241e-9b1ab829d419, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.202778] env[65385]: DEBUG nova.compute.manager [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 675.202778] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 675.203560] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa2119b3-bf8a-4dee-83ac-880580c7feda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.215639] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 675.216275] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2795ee9-6f10-4e92-8d54-a16e38250dc4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.230674] env[65385]: DEBUG oslo_vmware.api [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 675.230674] env[65385]: value = "task-4453291" [ 675.230674] env[65385]: _type = "Task" [ 675.230674] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.243126] env[65385]: DEBUG oslo_vmware.api [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453291, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.288042] env[65385]: INFO nova.compute.manager [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Took 31.75 seconds to build instance. [ 675.315256] env[65385]: WARNING openstack [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.316047] env[65385]: WARNING openstack [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.329532] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.401480] env[65385]: DEBUG oslo_vmware.api [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453290, 'name': PowerOffVM_Task, 'duration_secs': 0.449222} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.402152] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 675.402152] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 675.402286] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f1bcae2-33d7-4863-b1d3-f0f9596b8f11 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.479822] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 675.479960] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 675.480155] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Deleting the datastore file [datastore2] de8632e0-bf38-440d-b6a3-895efeef122d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 675.480435] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1228a362-d70b-418e-812b-cda931b30e6f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.489382] env[65385]: DEBUG oslo_vmware.api [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for the task: (returnval){ [ 675.489382] env[65385]: value = "task-4453293" [ 675.489382] env[65385]: _type = "Task" [ 675.489382] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.504221] env[65385]: DEBUG oslo_vmware.api [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453293, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.616799] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bb12e-cb26-e965-241e-9b1ab829d419, 'name': SearchDatastore_Task, 'duration_secs': 0.014461} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.617017] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 675.617526] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 675.617526] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.617722] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.618103] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 675.618181] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d135226-5d4a-4ad4-813e-b9da4e098b93 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.628480] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 675.628671] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 675.629489] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-860dd77c-3d3b-419b-8208-ff417b97ff80 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.635368] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf12ea8c-1d28-41de-b2ec-f2c1723b59e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.639690] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 675.639690] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b01663-b184-d756-71ef-a961defc2671" [ 675.639690] env[65385]: _type = "Task" [ 675.639690] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.646499] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829c8185-78e4-40a7-be50-adfcbd952d2e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.652738] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b01663-b184-d756-71ef-a961defc2671, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.682100] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5caf38a8-7452-4db3-819d-5355c9739603 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.690661] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb7d427-f024-48a9-9e5e-694fa6b53d83 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.705617] env[65385]: DEBUG nova.compute.provider_tree [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.743780] env[65385]: DEBUG oslo_vmware.api [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453291, 'name': PowerOffVM_Task, 'duration_secs': 0.294666} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.744220] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 675.744491] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 675.744802] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22318604-0f63-4b9c-b300-838596ab02dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.791313] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed46a53f-4de5-43c3-acc0-f7daea2c966f tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.260s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.817262] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 675.817777] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 675.817777] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Deleting the datastore file [datastore2] 69fdd5df-a9f2-486f-8a79-87c034366083 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 675.818285] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0045dd42-44a5-4e49-aa46-c48082320314 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.826113] env[65385]: DEBUG oslo_vmware.api [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for the task: (returnval){ [ 675.826113] env[65385]: value = "task-4453295" [ 675.826113] env[65385]: _type = "Task" [ 675.826113] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.835403] env[65385]: DEBUG oslo_vmware.api [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453295, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.006164] env[65385]: DEBUG oslo_vmware.api [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Task: {'id': task-4453293, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.365981} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.006543] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 676.006591] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 676.006733] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.006901] env[65385]: INFO nova.compute.manager [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 676.007158] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 676.007356] env[65385]: DEBUG nova.compute.manager [-] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 676.007453] env[65385]: DEBUG nova.network.neutron [-] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 676.007733] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.008276] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.008519] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.076677] env[65385]: WARNING neutronclient.v2_0.client [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.077267] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.077661] env[65385]: WARNING openstack [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.153044] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b01663-b184-d756-71ef-a961defc2671, 'name': SearchDatastore_Task, 'duration_secs': 0.014785} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.154248] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3baec6c2-f1fe-49ce-8bf5-32ba3616c1d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.162859] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 676.162859] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528a68a7-c968-bb9d-5253-d759aac75f2f" [ 676.162859] env[65385]: _type = "Task" [ 676.162859] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.176063] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528a68a7-c968-bb9d-5253-d759aac75f2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.209294] env[65385]: DEBUG nova.scheduler.client.report [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 676.218750] env[65385]: WARNING neutronclient.v2_0.client [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.219808] env[65385]: WARNING openstack [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.219808] env[65385]: WARNING openstack [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.268823] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.340053] env[65385]: DEBUG oslo_vmware.api [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Task: {'id': task-4453295, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189065} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.341332] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 676.341332] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 676.341332] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.341332] env[65385]: INFO nova.compute.manager [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Took 1.14 seconds to destroy the instance on the hypervisor. [ 676.341745] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 676.342082] env[65385]: DEBUG nova.compute.manager [-] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 676.342293] env[65385]: DEBUG nova.network.neutron [-] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 676.342710] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.343433] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.343928] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.429025] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Acquiring lock "8ac44124-d8cf-421c-9ae4-c943df94550a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.429303] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.429447] env[65385]: INFO nova.compute.manager [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Attaching volume 3b3ed2bd-062e-4d87-8958-ddafcb01885d to /dev/sdb [ 676.499065] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2527aff9-9788-4126-ab06-59dcad24efaa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.508592] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4441fc-481c-4f28-b4d2-33b7e85e918f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.521691] env[65385]: DEBUG nova.virt.block_device [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Updating existing volume attachment record: ecf63ef0-ed19-495f-9e52-7a68550ac0a9 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 676.680277] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528a68a7-c968-bb9d-5253-d759aac75f2f, 'name': SearchDatastore_Task, 'duration_secs': 0.042545} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.680277] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.680277] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 950efcc6-7e4f-4b30-b0fa-d940f893e1d5/950efcc6-7e4f-4b30-b0fa-d940f893e1d5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 676.680277] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1945adc-554b-4a44-9e95-d334e86770af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.690364] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 676.690364] env[65385]: value = "task-4453296" [ 676.690364] env[65385]: _type = "Task" [ 676.690364] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.701642] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453296, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.718272] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.718272] env[65385]: DEBUG nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 676.723549] env[65385]: DEBUG nova.network.neutron [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Updating instance_info_cache with network_info: [{"id": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "address": "fa:16:3e:9d:e2:9e", "network": {"id": "a97d30fb-abe7-4ead-a564-a64f533b6756", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1681302515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8167324e3664a578a5ca8c8a04c1241", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "537e0890-4fa2-4f2d-b74c-49933a4edf53", "external-id": "nsx-vlan-transportzone-82", "segmentation_id": 82, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd970616a-b8", "ovs_interfaceid": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 676.725394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.901s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.727634] env[65385]: INFO nova.compute.claims [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.789815] env[65385]: DEBUG nova.network.neutron [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Updated VIF entry in instance network info cache for port 07665fa0-2bf4-4c1d-9058-505e323caa56. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 676.790181] env[65385]: DEBUG nova.network.neutron [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Updating instance_info_cache with network_info: [{"id": "07665fa0-2bf4-4c1d-9058-505e323caa56", "address": "fa:16:3e:cd:c9:0c", "network": {"id": "1b05f7a0-d3eb-463c-bdfa-e1462758dce6", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1282242937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec2768a4aefb4e1b9484fb266d45d2a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07665fa0-2b", "ovs_interfaceid": "07665fa0-2bf4-4c1d-9058-505e323caa56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 676.811121] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.180380] env[65385]: DEBUG nova.network.neutron [-] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 677.203486] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453296, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.228161] env[65385]: DEBUG nova.compute.utils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 677.230456] env[65385]: DEBUG nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 677.237319] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Releasing lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.237319] env[65385]: DEBUG nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Instance network_info: |[{"id": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "address": "fa:16:3e:9d:e2:9e", "network": {"id": "a97d30fb-abe7-4ead-a564-a64f533b6756", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1681302515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8167324e3664a578a5ca8c8a04c1241", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "537e0890-4fa2-4f2d-b74c-49933a4edf53", "external-id": "nsx-vlan-transportzone-82", "segmentation_id": 82, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd970616a-b8", "ovs_interfaceid": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 677.237319] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:e2:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '537e0890-4fa2-4f2d-b74c-49933a4edf53', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd970616a-b8e1-4ab4-9f2e-e7b1f97dfff7', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 677.248691] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Creating folder: Project (c8167324e3664a578a5ca8c8a04c1241). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 677.250021] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db1c122a-1e4c-4aba-bfcf-56d57ae9a37e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.268145] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Created folder: Project (c8167324e3664a578a5ca8c8a04c1241) in parent group-v870881. [ 677.268145] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Creating folder: Instances. Parent ref: group-v870960. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 677.268145] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aac0bac8-135f-4bf2-a53c-d55a6534f1ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.282634] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Created folder: Instances in parent group-v870960. [ 677.283322] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 677.283322] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 677.283602] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd609c42-580c-4fbc-865a-c45c57349661 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.301626] env[65385]: DEBUG oslo_concurrency.lockutils [req-a6371ae0-6c59-45ae-94ed-6123d475037d req-f2f17f81-6b29-4228-8220-eaa5d7ff4dd1 service nova] Releasing lock "refresh_cache-950efcc6-7e4f-4b30-b0fa-d940f893e1d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.311657] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 677.311657] env[65385]: value = "task-4453303" [ 677.311657] env[65385]: _type = "Task" [ 677.311657] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.327148] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453303, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.684801] env[65385]: INFO nova.compute.manager [-] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Took 1.68 seconds to deallocate network for instance. [ 677.705844] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453296, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595388} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.707557] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 950efcc6-7e4f-4b30-b0fa-d940f893e1d5/950efcc6-7e4f-4b30-b0fa-d940f893e1d5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 677.707933] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 677.708289] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3efe9f20-2b56-4180-89f5-b4dff35572fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.717555] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 677.717555] env[65385]: value = "task-4453304" [ 677.717555] env[65385]: _type = "Task" [ 677.717555] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.730870] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453304, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.731892] env[65385]: DEBUG nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 677.812316] env[65385]: DEBUG nova.network.neutron [-] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 677.830040] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453303, 'name': CreateVM_Task, 'duration_secs': 0.492021} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.830214] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 677.830718] env[65385]: WARNING neutronclient.v2_0.client [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.831119] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.831268] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.831610] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 677.832391] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87e72672-4a56-4f47-856d-d0c7e3dc3918 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.838864] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 677.838864] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b35063-82ef-f836-ee30-f07f10b94510" [ 677.838864] env[65385]: _type = "Task" [ 677.838864] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.851491] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b35063-82ef-f836-ee30-f07f10b94510, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.191057] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748e1610-767d-4f3c-b72a-b7a718395d31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.196826] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.200744] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a419650-288f-4acb-ba08-601ebb197283 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.249343] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfc219e-798d-4939-ba02-222d8a5ae930 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.263378] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453304, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076877} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.263731] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 678.266344] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815d5a15-7ca0-4bdd-b49c-ab91dc357d08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.269707] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b0c002-b954-4990-91f3-c767ed890334 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.288030] env[65385]: DEBUG nova.compute.provider_tree [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 678.308239] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] 950efcc6-7e4f-4b30-b0fa-d940f893e1d5/950efcc6-7e4f-4b30-b0fa-d940f893e1d5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 678.313500] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-160bfede-fd20-4af1-a24c-c993354aeaf1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.329653] env[65385]: INFO nova.compute.manager [-] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Took 1.99 seconds to deallocate network for instance. [ 678.352199] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 678.352199] env[65385]: value = "task-4453305" [ 678.352199] env[65385]: _type = "Task" [ 678.352199] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.362563] env[65385]: ERROR nova.scheduler.client.report [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [req-eb2b7824-2b58-48d0-8840-9e02bd5aaf0f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-eb2b7824-2b58-48d0-8840-9e02bd5aaf0f"}]} [ 678.363651] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b35063-82ef-f836-ee30-f07f10b94510, 'name': SearchDatastore_Task, 'duration_secs': 0.01068} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.368956] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.368956] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 678.368956] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.368956] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 678.368956] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 678.369358] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3dc3707-12ec-475b-82e6-776dbaa714c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.376476] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.385790] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 678.386122] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 678.387125] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ce39800-9bb7-48f1-bd45-3e29aa481830 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.395024] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 678.395024] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5219f6a8-826e-572a-0d03-b4935995a740" [ 678.395024] env[65385]: _type = "Task" [ 678.395024] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.396274] env[65385]: DEBUG nova.scheduler.client.report [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 678.410731] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5219f6a8-826e-572a-0d03-b4935995a740, 'name': SearchDatastore_Task, 'duration_secs': 0.011596} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.411698] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80a2ad9a-bdc2-433c-942d-09417b7099d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.416747] env[65385]: DEBUG nova.scheduler.client.report [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 678.416997] env[65385]: DEBUG nova.compute.provider_tree [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 678.420549] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 678.420549] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e3d7ef-0607-541f-868c-9830f506e34b" [ 678.420549] env[65385]: _type = "Task" [ 678.420549] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.431625] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e3d7ef-0607-541f-868c-9830f506e34b, 'name': SearchDatastore_Task} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.433416] env[65385]: DEBUG nova.scheduler.client.report [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: ebd65cec-2677-410c-a4fa-45c33de4b6f1 {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 678.434788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.435096] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 2e4c064f-27b0-4ccb-8fe6-c34a61153663/2e4c064f-27b0-4ccb-8fe6-c34a61153663.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 678.435626] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3220a75e-6816-474c-8144-3cb70f12300b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.443952] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 678.443952] env[65385]: value = "task-4453306" [ 678.443952] env[65385]: _type = "Task" [ 678.443952] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.455413] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453306, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.455731] env[65385]: DEBUG nova.scheduler.client.report [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 678.754721] env[65385]: DEBUG nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 678.783892] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 678.784199] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 678.784388] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 678.784606] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 678.784756] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 678.784994] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 678.785109] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 678.785293] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 678.785433] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 678.785587] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 678.785752] env[65385]: DEBUG nova.virt.hardware [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 678.786654] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bfaf92-f442-45a6-b7b3-4c74b4f3f29f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.804517] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395ca4c4-c342-4571-92c1-f92f819fa94d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.822472] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.828816] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Creating folder: Project (0e778e71f7644764b8d0d87d973280eb). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.831935] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91421a46-b5c4-4664-a969-228fe3b7e3a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.845286] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Created folder: Project (0e778e71f7644764b8d0d87d973280eb) in parent group-v870881. [ 678.845483] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Creating folder: Instances. Parent ref: group-v870963. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.846488] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.846742] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a846e8f-31ba-49ad-8d43-1ea9989794cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.863632] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453305, 'name': ReconfigVM_Task, 'duration_secs': 0.426382} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.864118] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Reconfigured VM instance instance-0000001b to attach disk [datastore1] 950efcc6-7e4f-4b30-b0fa-d940f893e1d5/950efcc6-7e4f-4b30-b0fa-d940f893e1d5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 678.866111] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11824e45-3dea-441c-add9-c9ea6bb64eb9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.868341] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Created folder: Instances in parent group-v870963. [ 678.868635] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 678.868870] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 678.869418] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a6082859-41d1-481a-9d68-1696ac0d2a72 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.886054] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 678.886054] env[65385]: value = "task-4453309" [ 678.886054] env[65385]: _type = "Task" [ 678.886054] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.894039] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.894039] env[65385]: value = "task-4453310" [ 678.894039] env[65385]: _type = "Task" [ 678.894039] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.897855] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453309, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.910807] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453310, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.958614] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453306, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.972512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23f60a7-dfae-4990-b95a-be864a21a926 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.984618] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8cea74-58d8-453b-b976-ffec0dbd2d2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.023071] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334bba52-5b6b-4fa4-993a-7e1b7b70ed20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.034207] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbaa8cca-9184-4fbd-be73-ff95537306a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.056357] env[65385]: DEBUG nova.compute.provider_tree [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 679.298967] env[65385]: DEBUG nova.compute.manager [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 679.299922] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31aa318c-8437-42da-b76b-551413fc0a0d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.398281] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453309, 'name': Rename_Task, 'duration_secs': 0.400184} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.401609] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 679.402330] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c385a65-a525-43f5-93df-9fb71cb0db43 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.409725] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453310, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.411135] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 679.411135] env[65385]: value = "task-4453312" [ 679.411135] env[65385]: _type = "Task" [ 679.411135] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.419960] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453312, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.457929] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453306, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660736} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.458363] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 2e4c064f-27b0-4ccb-8fe6-c34a61153663/2e4c064f-27b0-4ccb-8fe6-c34a61153663.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 679.458621] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 679.458918] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f370c27c-41b0-479b-b6d2-0417568a820a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.467767] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 679.467767] env[65385]: value = "task-4453313" [ 679.467767] env[65385]: _type = "Task" [ 679.467767] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.478776] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453313, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.588873] env[65385]: DEBUG nova.compute.manager [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Received event network-vif-plugged-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 679.589530] env[65385]: DEBUG oslo_concurrency.lockutils [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Acquiring lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 679.590025] env[65385]: DEBUG oslo_concurrency.lockutils [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 679.594022] env[65385]: DEBUG oslo_concurrency.lockutils [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.594022] env[65385]: DEBUG nova.compute.manager [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] No waiting events found dispatching network-vif-plugged-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 679.594022] env[65385]: WARNING nova.compute.manager [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Received unexpected event network-vif-plugged-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 for instance with vm_state building and task_state spawning. [ 679.594022] env[65385]: DEBUG nova.compute.manager [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Received event network-changed-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 679.594022] env[65385]: DEBUG nova.compute.manager [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Refreshing instance network info cache due to event network-changed-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 679.594022] env[65385]: DEBUG oslo_concurrency.lockutils [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Acquiring lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.594022] env[65385]: DEBUG oslo_concurrency.lockutils [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Acquired lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.594022] env[65385]: DEBUG nova.network.neutron [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Refreshing network info cache for port d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 679.620698] env[65385]: DEBUG nova.scheduler.client.report [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 44 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 679.620956] env[65385]: DEBUG nova.compute.provider_tree [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 44 to 45 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 679.621144] env[65385]: DEBUG nova.compute.provider_tree [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 679.733906] env[65385]: DEBUG nova.compute.manager [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Received event network-vif-deleted-7d3f8bf5-22d2-41b0-9743-0616daaef615 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 679.734115] env[65385]: DEBUG nova.compute.manager [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Received event network-changed-18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 679.734485] env[65385]: DEBUG nova.compute.manager [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Refreshing instance network info cache due to event network-changed-18c18d61-0493-40e3-8883-c90faabf147b. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 679.734737] env[65385]: DEBUG oslo_concurrency.lockutils [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Acquiring lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.735037] env[65385]: DEBUG oslo_concurrency.lockutils [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Acquired lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.735095] env[65385]: DEBUG nova.network.neutron [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Refreshing network info cache for port 18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 679.815396] env[65385]: INFO nova.compute.manager [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] instance snapshotting [ 679.817227] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac66137a-79dc-4cb8-90c6-b78bd03d839c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.844409] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75c2e82-a91b-4df0-b2f6-96316e22c130 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.914007] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453310, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.924558] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453312, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.982078] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453313, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128234} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.982486] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 679.984898] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8e664c-3d66-4835-b7b7-1e2c8e832f21 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.013655] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 2e4c064f-27b0-4ccb-8fe6-c34a61153663/2e4c064f-27b0-4ccb-8fe6-c34a61153663.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 680.014144] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbd73cd2-c57b-4c89-b9b3-5e5afdbdeb79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.037198] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 680.037198] env[65385]: value = "task-4453314" [ 680.037198] env[65385]: _type = "Task" [ 680.037198] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.050344] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453314, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.096272] env[65385]: WARNING neutronclient.v2_0.client [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.097182] env[65385]: WARNING openstack [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.097578] env[65385]: WARNING openstack [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.127514] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.402s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.129425] env[65385]: DEBUG nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 680.132513] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.399s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.134666] env[65385]: INFO nova.compute.claims [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.239777] env[65385]: WARNING neutronclient.v2_0.client [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.240523] env[65385]: WARNING openstack [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.240834] env[65385]: WARNING openstack [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.363846] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 680.364586] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "a71d2fa6-006c-4f76-8ae8-467e149297da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.364586] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.364909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "a71d2fa6-006c-4f76-8ae8-467e149297da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.364909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.365045] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.372404] env[65385]: WARNING openstack [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.372873] env[65385]: WARNING openstack [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.383426] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-43687e9e-3ada-42c2-b702-554e6b717294 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.387821] env[65385]: INFO nova.compute.manager [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Terminating instance [ 680.400660] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 680.400660] env[65385]: value = "task-4453315" [ 680.400660] env[65385]: _type = "Task" [ 680.400660] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.420482] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453315, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.420629] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453310, 'name': CreateVM_Task, 'duration_secs': 1.395908} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.427702] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 680.428476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.428647] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.428990] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 680.429650] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5aa65832-4443-4379-9f3a-4d8d1f577c7e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.435403] env[65385]: DEBUG oslo_vmware.api [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453312, 'name': PowerOnVM_Task, 'duration_secs': 0.625527} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.437179] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 680.437563] env[65385]: INFO nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Took 10.54 seconds to spawn the instance on the hypervisor. [ 680.437563] env[65385]: DEBUG nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 680.438272] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 680.438272] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fc5fdf-8fe3-95f8-9472-0f7bc336dcf2" [ 680.438272] env[65385]: _type = "Task" [ 680.438272] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.439029] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c419c65-e88f-449c-b437-f95e0651c818 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.460354] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fc5fdf-8fe3-95f8-9472-0f7bc336dcf2, 'name': SearchDatastore_Task, 'duration_secs': 0.013035} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.461114] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.461114] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 680.461334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.461468] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.461721] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 680.462820] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4acad37a-f8bf-4159-b0f1-e53d4f1103e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.475738] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 680.475902] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 680.476969] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3941a4e5-e28e-4fe4-a679-6bdd4ed419d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.484857] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 680.484857] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522345b1-ab42-b82c-5217-fcee458c5b5e" [ 680.484857] env[65385]: _type = "Task" [ 680.484857] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.494873] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522345b1-ab42-b82c-5217-fcee458c5b5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.553145] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453314, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.569968] env[65385]: WARNING openstack [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.570490] env[65385]: WARNING openstack [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.609905] env[65385]: WARNING neutronclient.v2_0.client [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.611244] env[65385]: WARNING openstack [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.611244] env[65385]: WARNING openstack [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.647019] env[65385]: DEBUG nova.compute.utils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 680.649056] env[65385]: DEBUG nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 680.649484] env[65385]: DEBUG nova.network.neutron [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 680.650312] env[65385]: WARNING neutronclient.v2_0.client [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.650312] env[65385]: WARNING neutronclient.v2_0.client [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.650595] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.651082] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.664071] env[65385]: DEBUG nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 680.711558] env[65385]: WARNING neutronclient.v2_0.client [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.712224] env[65385]: WARNING openstack [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.712593] env[65385]: WARNING openstack [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.754971] env[65385]: DEBUG nova.policy [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2a287457352470794887b229ea90cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60de0d1c162342209795a180391661f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 680.853473] env[65385]: DEBUG nova.network.neutron [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Updated VIF entry in instance network info cache for port d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 680.853838] env[65385]: DEBUG nova.network.neutron [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Updating instance_info_cache with network_info: [{"id": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "address": "fa:16:3e:9d:e2:9e", "network": {"id": "a97d30fb-abe7-4ead-a564-a64f533b6756", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1681302515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8167324e3664a578a5ca8c8a04c1241", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "537e0890-4fa2-4f2d-b74c-49933a4edf53", "external-id": "nsx-vlan-transportzone-82", "segmentation_id": 82, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd970616a-b8", "ovs_interfaceid": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.893998] env[65385]: DEBUG nova.compute.manager [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 680.894466] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 680.895650] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34c85a6-c5b4-4e22-9f39-33fb663f9921 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.922287] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 680.923820] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-937c4fda-e44b-4e77-bc51-27333b0d69a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.931592] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453315, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.940365] env[65385]: DEBUG oslo_vmware.api [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 680.940365] env[65385]: value = "task-4453316" [ 680.940365] env[65385]: _type = "Task" [ 680.940365] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.951412] env[65385]: DEBUG oslo_vmware.api [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453316, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.970012] env[65385]: DEBUG nova.network.neutron [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updated VIF entry in instance network info cache for port 18c18d61-0493-40e3-8883-c90faabf147b. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 680.970459] env[65385]: DEBUG nova.network.neutron [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.981447] env[65385]: INFO nova.compute.manager [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Took 35.36 seconds to build instance. [ 681.003384] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522345b1-ab42-b82c-5217-fcee458c5b5e, 'name': SearchDatastore_Task, 'duration_secs': 0.017719} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.004905] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1809d31-09bb-4e45-b76c-d96c514d49c5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.012704] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 681.012704] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284751c-0e5f-9301-4479-b4e616652a05" [ 681.012704] env[65385]: _type = "Task" [ 681.012704] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.033766] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284751c-0e5f-9301-4479-b4e616652a05, 'name': SearchDatastore_Task, 'duration_secs': 0.012932} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.035910] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.035910] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 681.035910] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-211cb726-b035-4fb7-b4e5-2ed5e96c78b3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.045869] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 681.045869] env[65385]: value = "task-4453317" [ 681.045869] env[65385]: _type = "Task" [ 681.045869] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.056903] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453314, 'name': ReconfigVM_Task, 'duration_secs': 0.64938} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.056903] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 2e4c064f-27b0-4ccb-8fe6-c34a61153663/2e4c064f-27b0-4ccb-8fe6-c34a61153663.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 681.056903] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf58ec2c-5db0-41ac-9aff-83b32295f63f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.063410] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.070728] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "2f66a03b-874b-4bee-9694-49813ecb0c8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.070959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.072320] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "90c36f67-0c25-4487-a5e4-89827290953b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.072378] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "90c36f67-0c25-4487-a5e4-89827290953b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.073470] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 681.073470] env[65385]: value = "task-4453318" [ 681.073470] env[65385]: _type = "Task" [ 681.073470] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.086134] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453318, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.119669] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 681.120162] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870959', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'name': 'volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ac44124-d8cf-421c-9ae4-c943df94550a', 'attached_at': '', 'detached_at': '', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'serial': '3b3ed2bd-062e-4d87-8958-ddafcb01885d'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 681.121382] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee982a23-801d-44e2-8f68-da29fc6ab5cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.144515] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85fd292-bcb0-4a88-b6b0-f850b4d034f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.176730] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d/volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 681.181578] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa88d9cb-59ed-4cec-b13c-e463860f1e77 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.208037] env[65385]: DEBUG oslo_vmware.api [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Waiting for the task: (returnval){ [ 681.208037] env[65385]: value = "task-4453319" [ 681.208037] env[65385]: _type = "Task" [ 681.208037] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.219801] env[65385]: DEBUG oslo_vmware.api [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453319, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.361851] env[65385]: DEBUG oslo_concurrency.lockutils [req-a0e2195c-7239-4c37-8df9-c39ecf1d10e5 req-dbbb70ac-fc07-46af-91cc-f4480f50ee50 service nova] Releasing lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.416978] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.417398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.417602] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.417832] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.417985] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.429256] env[65385]: INFO nova.compute.manager [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Terminating instance [ 681.442291] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453315, 'name': CreateSnapshot_Task, 'duration_secs': 0.930502} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.446316] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 681.448117] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548c12df-bedd-430f-b1d4-5b190e4539e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.468214] env[65385]: DEBUG oslo_vmware.api [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453316, 'name': PowerOffVM_Task, 'duration_secs': 0.232094} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.472327] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 681.472611] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 681.473294] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4ea7f5a-db0d-4d16-a874-ee0a82d62977 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.475943] env[65385]: DEBUG oslo_concurrency.lockutils [req-dc7b84c2-c9f7-4a03-a487-b237990b8d9f req-7efd6dfb-f84c-4359-b00d-93b1fca53614 service nova] Releasing lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.484932] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6a42c60b-09fb-492b-b981-81da46ff8518 tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.888s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.501108] env[65385]: DEBUG nova.network.neutron [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Successfully created port: 59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 681.564950] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453317, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.569911] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 681.570164] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 681.570339] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Deleting the datastore file [datastore1] a71d2fa6-006c-4f76-8ae8-467e149297da {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 681.570625] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2843e536-663b-4f8a-aee5-c737a1a9288d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.576420] env[65385]: DEBUG nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 681.584593] env[65385]: DEBUG nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 681.591907] env[65385]: DEBUG oslo_vmware.api [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for the task: (returnval){ [ 681.591907] env[65385]: value = "task-4453321" [ 681.591907] env[65385]: _type = "Task" [ 681.591907] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.603083] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453318, 'name': Rename_Task, 'duration_secs': 0.180148} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.604427] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 681.604824] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5919e443-f9ae-45af-aef6-45810a2080e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.611418] env[65385]: DEBUG oslo_vmware.api [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.620332] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 681.620332] env[65385]: value = "task-4453322" [ 681.620332] env[65385]: _type = "Task" [ 681.620332] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.632491] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453322, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.700674] env[65385]: DEBUG nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 681.719815] env[65385]: DEBUG oslo_vmware.api [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453319, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.734246] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 681.734505] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 681.734717] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 681.734891] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 681.735078] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 681.735235] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 681.735434] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.735712] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 681.735910] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 681.736077] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 681.736280] env[65385]: DEBUG nova.virt.hardware [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 681.738276] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a83b59-c569-4572-b3be-bfa875749b02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.749453] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f0c253-f667-480e-9779-7ca5039f1971 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.856395] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0162b827-5cef-45d1-9139-19808d96656a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.866531] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265e949d-87a5-4e0b-a43e-12e8ea789fb6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.900662] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d560d6-b887-4382-9601-57e96215a993 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.909060] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41464cf3-94e1-4db3-b44d-88ea83893653 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.924442] env[65385]: DEBUG nova.compute.provider_tree [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.934728] env[65385]: DEBUG nova.compute.manager [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 681.935252] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.935819] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6fbaa9-60d8-4241-995a-188fb7c34f91 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.945365] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 681.945637] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d4559ce-0b4c-4ed9-a826-00ce50f93a27 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.954488] env[65385]: DEBUG oslo_vmware.api [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 681.954488] env[65385]: value = "task-4453323" [ 681.954488] env[65385]: _type = "Task" [ 681.954488] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.963158] env[65385]: DEBUG oslo_vmware.api [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.980898] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 681.981246] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-637a2e65-001d-4312-a246-2affb14d4687 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.990592] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 681.990592] env[65385]: value = "task-4453324" [ 681.990592] env[65385]: _type = "Task" [ 681.990592] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.004170] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453324, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.060922] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453317, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608773} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.061162] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 682.062104] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 682.062104] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7803853c-6923-423b-8659-d1291e8bd6c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.071762] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 682.071762] env[65385]: value = "task-4453325" [ 682.071762] env[65385]: _type = "Task" [ 682.071762] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.083211] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453325, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.110894] env[65385]: DEBUG oslo_vmware.api [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.115808] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.117066] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.132177] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453322, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.228170] env[65385]: DEBUG oslo_vmware.api [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453319, 'name': ReconfigVM_Task, 'duration_secs': 0.518522} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.228170] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Reconfigured VM instance instance-0000000a to attach disk [datastore1] volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d/volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 682.234254] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b50d924f-cb34-4748-923e-1d8ac7b2955b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.255343] env[65385]: DEBUG oslo_vmware.api [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Waiting for the task: (returnval){ [ 682.255343] env[65385]: value = "task-4453326" [ 682.255343] env[65385]: _type = "Task" [ 682.255343] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.268731] env[65385]: DEBUG oslo_vmware.api [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453326, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.429345] env[65385]: DEBUG nova.scheduler.client.report [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 682.466520] env[65385]: DEBUG oslo_vmware.api [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453323, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.504106] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453324, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.586702] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453325, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.25569} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.587189] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.589486] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351eae93-988a-46e9-b7e0-a0bffa9a4833 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.621850] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.622292] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2228c40d-a50c-4d11-8e13-fba79e357db9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.650144] env[65385]: DEBUG oslo_vmware.api [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Task: {'id': task-4453321, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.532267} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.650955] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 682.651247] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 682.651481] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 682.651752] env[65385]: INFO nova.compute.manager [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Took 1.76 seconds to destroy the instance on the hypervisor. [ 682.652037] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 682.652330] env[65385]: DEBUG nova.compute.manager [-] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 682.652499] env[65385]: DEBUG nova.network.neutron [-] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 682.652816] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 682.653903] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.653989] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.668309] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 682.668309] env[65385]: value = "task-4453327" [ 682.668309] env[65385]: _type = "Task" [ 682.668309] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.668309] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453322, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.678031] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453327, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.751086] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 682.771366] env[65385]: DEBUG oslo_vmware.api [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453326, 'name': ReconfigVM_Task, 'duration_secs': 0.201134} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.771987] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870959', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'name': 'volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ac44124-d8cf-421c-9ae4-c943df94550a', 'attached_at': '', 'detached_at': '', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'serial': '3b3ed2bd-062e-4d87-8958-ddafcb01885d'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 682.937089] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.804s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.937695] env[65385]: DEBUG nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 682.940982] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.447s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.942693] env[65385]: INFO nova.compute.claims [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.965692] env[65385]: DEBUG oslo_vmware.api [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453323, 'name': PowerOffVM_Task, 'duration_secs': 0.526061} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.966821] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 682.966821] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 682.966821] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c3ec2ec-e36d-43f5-9538-1e3e85817708 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.003703] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453324, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.138370] env[65385]: DEBUG oslo_vmware.api [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453322, 'name': PowerOnVM_Task, 'duration_secs': 1.305802} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.138768] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 683.139085] env[65385]: INFO nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Took 11.10 seconds to spawn the instance on the hypervisor. [ 683.139468] env[65385]: DEBUG nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 683.140237] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0447f607-c73c-41a3-adf1-61c7e6b51a3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.154541] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 683.154541] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 683.154541] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Deleting the datastore file [datastore2] 5291c04c-24ad-4c64-9fc2-fde42da4bc0f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 683.155306] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d9dbaa7-2044-4da2-b93b-c84585df9df9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.163851] env[65385]: DEBUG oslo_vmware.api [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 683.163851] env[65385]: value = "task-4453329" [ 683.163851] env[65385]: _type = "Task" [ 683.163851] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.175747] env[65385]: DEBUG oslo_vmware.api [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.180542] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453327, 'name': ReconfigVM_Task, 'duration_secs': 0.402957} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.180802] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Reconfigured VM instance instance-0000001d to attach disk [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 683.181448] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c899d40e-58f2-438f-a349-367794c2bd4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.192750] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 683.192750] env[65385]: value = "task-4453330" [ 683.192750] env[65385]: _type = "Task" [ 683.192750] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.204054] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453330, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.371283] env[65385]: DEBUG nova.network.neutron [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Successfully updated port: 59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 683.448168] env[65385]: DEBUG nova.compute.utils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 683.454907] env[65385]: DEBUG nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 683.454907] env[65385]: DEBUG nova.network.neutron [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 683.454907] env[65385]: WARNING neutronclient.v2_0.client [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.454907] env[65385]: WARNING neutronclient.v2_0.client [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.455402] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.455809] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.503628] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453324, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.535836] env[65385]: DEBUG nova.policy [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ba4783b423a4120a1035fabb02cd0aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9a1bf7c2dad4808a3782a3fba8a25fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 683.603084] env[65385]: DEBUG nova.network.neutron [-] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 683.674816] env[65385]: INFO nova.compute.manager [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Took 35.98 seconds to build instance. [ 683.682416] env[65385]: DEBUG oslo_vmware.api [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.372884} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.682673] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 683.683028] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 683.683028] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.683208] env[65385]: INFO nova.compute.manager [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Took 1.75 seconds to destroy the instance on the hypervisor. [ 683.683979] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 683.683979] env[65385]: DEBUG nova.compute.manager [-] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 683.683979] env[65385]: DEBUG nova.network.neutron [-] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 683.683979] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.684536] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.685226] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.704860] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453330, 'name': Rename_Task, 'duration_secs': 0.179021} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.705606] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 683.705606] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b527ecd-4108-4a00-902d-dfd08e56a087 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.713057] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 683.713057] env[65385]: value = "task-4453331" [ 683.713057] env[65385]: _type = "Task" [ 683.713057] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.724258] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453331, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.858082] env[65385]: DEBUG nova.objects.instance [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lazy-loading 'flavor' on Instance uuid 8ac44124-d8cf-421c-9ae4-c943df94550a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 683.860410] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.874995] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.874995] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.874995] env[65385]: DEBUG nova.network.neutron [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 683.953376] env[65385]: DEBUG nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 683.963265] env[65385]: DEBUG nova.network.neutron [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Successfully created port: 29fa73f0-74dc-443a-b870-f4683fdc5d74 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 684.013959] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453324, 'name': CloneVM_Task, 'duration_secs': 1.648141} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.014324] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Created linked-clone VM from snapshot [ 684.015251] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686b4056-bb4e-49f1-b895-c5fc554287a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.029249] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Uploading image 10940f07-a2d5-4974-ab20-4280c29c352e {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 684.049674] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 684.050171] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-eb234773-3b87-4cbe-8a99-4bc16b2690b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.060320] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 684.060320] env[65385]: value = "task-4453332" [ 684.060320] env[65385]: _type = "Task" [ 684.060320] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.073489] env[65385]: DEBUG nova.compute.manager [req-3055570f-1a21-4518-a161-df1057ebb9ac req-0c25580b-8440-49fa-ab6e-d55a266a5c7e service nova] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Received event network-vif-deleted-e6913ace-1d77-409d-8108-5f01780b3051 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 684.075041] env[65385]: DEBUG nova.compute.manager [req-3055570f-1a21-4518-a161-df1057ebb9ac req-0c25580b-8440-49fa-ab6e-d55a266a5c7e service nova] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Received event network-vif-deleted-7454821e-9f2a-416b-a1ff-af59335a59a9 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 684.083908] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453332, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.107102] env[65385]: INFO nova.compute.manager [-] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Took 1.45 seconds to deallocate network for instance. [ 684.177323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6be30bda-4ade-41ce-a13a-350069f76905 tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.946s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.203507] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "483cf486-f0fc-48a4-9db2-970d0f35865a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.203961] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.214221] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "7fcbdc79-688c-479f-94e2-f4542abe714b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.214221] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.241791] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453331, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.362854] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6867c94e-65a2-4fa2-918b-730c00ef08b8 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.933s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.378307] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.378801] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.570672] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453332, 'name': Destroy_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.617238] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.649397] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd8ecbc-b2bf-40e6-8489-dec817844e01 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.659419] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b24772-ad03-4329-aab0-fd431c8266bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.698071] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c0396f-65b3-4c7e-857c-b46f00cd4150 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.701629] env[65385]: DEBUG nova.network.neutron [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 684.706656] env[65385]: DEBUG nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 684.713412] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa401ee-ae35-46f7-a74a-cb3091d0a6ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.732030] env[65385]: DEBUG nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 684.734500] env[65385]: DEBUG nova.compute.provider_tree [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.738815] env[65385]: DEBUG oslo_vmware.api [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453331, 'name': PowerOnVM_Task, 'duration_secs': 0.726274} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.739292] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 684.743029] env[65385]: INFO nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Took 5.98 seconds to spawn the instance on the hypervisor. [ 684.743029] env[65385]: DEBUG nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 684.743029] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d2aee1-ad5e-43fe-9a8a-3d72b2123288 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.963789] env[65385]: DEBUG nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 684.992267] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 684.992629] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 684.992803] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 684.993029] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 684.993212] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 684.993403] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 684.993653] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.993843] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 684.994052] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 684.994253] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 684.994660] env[65385]: DEBUG nova.virt.hardware [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 684.995424] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7b7b8f-b387-4f21-8d96-745ddefc7387 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.004391] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e10e54-46a0-4052-9a20-21ffdec2cd67 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.074323] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453332, 'name': Destroy_Task, 'duration_secs': 0.517165} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.076707] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Destroyed the VM [ 685.077328] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 685.077966] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-59093c40-ce1f-4af2-8edc-270dbf13bd2c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.090473] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 685.090473] env[65385]: value = "task-4453333" [ 685.090473] env[65385]: _type = "Task" [ 685.090473] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.105229] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453333, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.106716] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.107492] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.240686] env[65385]: DEBUG nova.scheduler.client.report [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 685.249274] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.264383] env[65385]: INFO nova.compute.manager [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Took 25.32 seconds to build instance. [ 685.289330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.309214] env[65385]: WARNING neutronclient.v2_0.client [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.311229] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.311924] env[65385]: WARNING openstack [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.338851] env[65385]: DEBUG nova.compute.manager [req-0737e746-ddb0-41a8-b9c5-470ae1c2df3c req-4d7c0071-a0bf-401c-bb64-4dc0ffb0e2a6 service nova] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Received event network-vif-deleted-8227e6e0-c35a-450c-8bf3-f9b038dd9116 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 685.579277] env[65385]: DEBUG nova.network.neutron [-] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 685.604272] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453333, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.668553] env[65385]: DEBUG nova.network.neutron [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 685.753859] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.813s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 685.756181] env[65385]: DEBUG nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 685.757758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.820s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 685.759270] env[65385]: INFO nova.compute.claims [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.766826] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f8063a95-2af5-4428-b06a-731e46855356 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.840s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 685.800622] env[65385]: DEBUG nova.network.neutron [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Successfully updated port: 29fa73f0-74dc-443a-b870-f4683fdc5d74 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 686.082120] env[65385]: INFO nova.compute.manager [-] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Took 2.40 seconds to deallocate network for instance. [ 686.104091] env[65385]: DEBUG oslo_vmware.api [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453333, 'name': RemoveSnapshot_Task, 'duration_secs': 0.778415} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.105228] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 686.174750] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 686.174750] env[65385]: DEBUG nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Instance network_info: |[{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 686.174750] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:a4:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59289962-8634-414f-9bd2-b3f5b19af9b4', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 686.183354] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Creating folder: Project (60de0d1c162342209795a180391661f6). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 686.184134] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c31d33c-4e52-4b55-9428-88e871e93bbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.197732] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Created folder: Project (60de0d1c162342209795a180391661f6) in parent group-v870881. [ 686.197965] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Creating folder: Instances. Parent ref: group-v870968. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 686.198204] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11fd9533-8311-4dfc-8e02-fc6ca73b7a4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.209073] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Created folder: Instances in parent group-v870968. [ 686.209404] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 686.211039] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 686.211039] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d55e1d0d-4334-4075-a1cf-e722ab48d33d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.238389] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 686.238389] env[65385]: value = "task-4453336" [ 686.238389] env[65385]: _type = "Task" [ 686.238389] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.248758] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453336, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.266734] env[65385]: DEBUG nova.compute.utils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 686.268717] env[65385]: DEBUG nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 686.268975] env[65385]: DEBUG nova.network.neutron [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 686.269361] env[65385]: WARNING neutronclient.v2_0.client [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.269738] env[65385]: WARNING neutronclient.v2_0.client [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.270896] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.271344] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.302491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "refresh_cache-0673a0d2-76ac-4cd5-8cc3-b8596877c641" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.302725] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "refresh_cache-0673a0d2-76ac-4cd5-8cc3-b8596877c641" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.302968] env[65385]: DEBUG nova.network.neutron [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 686.351707] env[65385]: DEBUG nova.policy [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0f069272314a31ad33c77137513a17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c27eb6232a2148a1a259f57494b4ae30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 686.591383] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.612760] env[65385]: WARNING nova.compute.manager [None req-50af5713-b0e9-478e-81f9-cf92db6ef31c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Image not found during snapshot: nova.exception.ImageNotFound: Image 10940f07-a2d5-4974-ab20-4280c29c352e could not be found. [ 686.725061] env[65385]: DEBUG nova.network.neutron [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Successfully created port: 6b67e7c9-05f5-47d9-9b81-86464d52033c {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 686.749769] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453336, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.754277] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "649b4e21-fc55-415c-a1f3-ec724397b874" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.754915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.754915] env[65385]: DEBUG nova.compute.manager [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 686.756895] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322bd9f7-ad5a-4688-9185-f362c3d7e649 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.761836] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.762421] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.762421] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.762529] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.763028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 686.766594] env[65385]: DEBUG nova.compute.manager [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 686.766594] env[65385]: DEBUG nova.objects.instance [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lazy-loading 'flavor' on Instance uuid 649b4e21-fc55-415c-a1f3-ec724397b874 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 686.768055] env[65385]: INFO nova.compute.manager [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Terminating instance [ 686.781144] env[65385]: DEBUG nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 686.809673] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.810358] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.220279] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce4510d-e426-491c-ae85-b82d1d894991 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.230456] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dfa620-6fc8-4269-9210-c39dac992050 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.264049] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a071392-dade-4c5f-a1e0-41e3248285ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.276141] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90316609-a21c-40a8-acf4-a019ba1dfaaf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.281839] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453336, 'name': CreateVM_Task, 'duration_secs': 0.597316} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.283356] env[65385]: DEBUG nova.compute.manager [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 687.283607] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.284078] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 687.284886] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2838b1c7-ac22-4144-aade-7d3702d5906e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.290486] env[65385]: WARNING neutronclient.v2_0.client [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.290876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.291078] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 687.291407] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 687.295369] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63a87e00-158e-4ba9-aeef-30d545d3882b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.306624] env[65385]: DEBUG nova.compute.provider_tree [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.311695] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 687.315032] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a68a3ec-15ff-4819-a082-324a208d3ffc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.316131] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 687.316131] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b327e-7c69-f4d8-5c86-3ac781b33bb2" [ 687.316131] env[65385]: _type = "Task" [ 687.316131] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.324558] env[65385]: DEBUG oslo_vmware.api [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 687.324558] env[65385]: value = "task-4453337" [ 687.324558] env[65385]: _type = "Task" [ 687.324558] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.328803] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b327e-7c69-f4d8-5c86-3ac781b33bb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.338672] env[65385]: DEBUG oslo_vmware.api [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453337, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.518856] env[65385]: DEBUG nova.network.neutron [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 687.568416] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Received event network-vif-plugged-59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 687.568642] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.568929] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 687.568988] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 687.569653] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] No waiting events found dispatching network-vif-plugged-59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 687.569653] env[65385]: WARNING nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Received unexpected event network-vif-plugged-59289962-8634-414f-9bd2-b3f5b19af9b4 for instance with vm_state building and task_state spawning. [ 687.569653] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Received event network-changed-59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 687.569653] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Refreshing instance network info cache due to event network-changed-59289962-8634-414f-9bd2-b3f5b19af9b4. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 687.570358] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.570358] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 687.570358] env[65385]: DEBUG nova.network.neutron [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Refreshing network info cache for port 59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 687.785404] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 687.785744] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b05fb885-0eb3-4317-92db-de9912df5b14 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.795414] env[65385]: DEBUG oslo_vmware.api [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 687.795414] env[65385]: value = "task-4453338" [ 687.795414] env[65385]: _type = "Task" [ 687.795414] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.805915] env[65385]: DEBUG oslo_vmware.api [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453338, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.807039] env[65385]: DEBUG nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 687.810344] env[65385]: DEBUG nova.scheduler.client.report [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.826933] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b327e-7c69-f4d8-5c86-3ac781b33bb2, 'name': SearchDatastore_Task, 'duration_secs': 0.018669} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.827267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 687.827496] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 687.827725] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.827866] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 687.828049] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 687.830377] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82de7214-d2cf-401a-a936-6952b14baef3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.839373] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 687.839614] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 687.839800] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 687.840212] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 687.840212] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 687.840212] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 687.840415] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.840567] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 687.840725] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 687.840884] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 687.841058] env[65385]: DEBUG nova.virt.hardware [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 687.842640] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a09a7ee-1da8-452b-9bf1-9f27b7a7e76a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.850453] env[65385]: DEBUG oslo_vmware.api [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453337, 'name': PowerOffVM_Task, 'duration_secs': 0.313056} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.850715] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 687.850872] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 687.852064] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 687.852272] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 687.852602] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bec1d933-b970-4c98-8581-d68474602962 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.857820] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b99ce530-5a3f-4285-b331-a6cf0b3b0007 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.861883] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9005765-d57c-466a-8b0e-f237d69bd270 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.868788] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 687.868788] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cae30d-a1c4-6903-b733-b053c1e936d8" [ 687.868788] env[65385]: _type = "Task" [ 687.868788] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.890204] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cae30d-a1c4-6903-b733-b053c1e936d8, 'name': SearchDatastore_Task, 'duration_secs': 0.016898} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.891119] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4238a6c8-44aa-4b97-a587-d89cd7ff9af4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.899183] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 687.899183] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5214c8d4-100d-6400-ef12-c82ec3d08d51" [ 687.899183] env[65385]: _type = "Task" [ 687.899183] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.908645] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5214c8d4-100d-6400-ef12-c82ec3d08d51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.945804] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 687.945804] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 687.945804] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Deleting the datastore file [datastore1] 950efcc6-7e4f-4b30-b0fa-d940f893e1d5 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 687.945804] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7064940-8f41-4565-a64c-1d5e39c5fbb0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.954444] env[65385]: DEBUG oslo_vmware.api [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for the task: (returnval){ [ 687.954444] env[65385]: value = "task-4453340" [ 687.954444] env[65385]: _type = "Task" [ 687.954444] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.965283] env[65385]: DEBUG oslo_vmware.api [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453340, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.074602] env[65385]: WARNING neutronclient.v2_0.client [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.074978] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.075471] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.255900] env[65385]: DEBUG nova.network.neutron [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Successfully updated port: 6b67e7c9-05f5-47d9-9b81-86464d52033c {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 688.306577] env[65385]: DEBUG oslo_vmware.api [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453338, 'name': PowerOffVM_Task, 'duration_secs': 0.22493} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.306860] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 688.307101] env[65385]: DEBUG nova.compute.manager [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 688.308029] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87b214d-4ccc-4a4a-a2cc-517e8b147472 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.315314] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.316044] env[65385]: DEBUG nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 688.321266] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 18.483s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 688.364946] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.365376] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.410981] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5214c8d4-100d-6400-ef12-c82ec3d08d51, 'name': SearchDatastore_Task, 'duration_secs': 0.023261} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.411181] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 688.411445] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 59f26752-75ab-45e5-b0a9-bbcef44ecd7f/59f26752-75ab-45e5-b0a9-bbcef44ecd7f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 688.416031] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd2960e3-7d77-43a4-91f6-da0b69c56149 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.430358] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 688.430358] env[65385]: value = "task-4453341" [ 688.430358] env[65385]: _type = "Task" [ 688.430358] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.446170] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.464430] env[65385]: DEBUG oslo_vmware.api [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Task: {'id': task-4453340, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159541} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.464699] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 688.464942] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 688.465072] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 688.465216] env[65385]: INFO nova.compute.manager [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Took 1.18 seconds to destroy the instance on the hypervisor. [ 688.465442] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 688.465627] env[65385]: DEBUG nova.compute.manager [-] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 688.465720] env[65385]: DEBUG nova.network.neutron [-] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 688.466053] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.466574] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.466958] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.652449] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.717676] env[65385]: WARNING neutronclient.v2_0.client [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.718358] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.718697] env[65385]: WARNING openstack [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.759045] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "refresh_cache-54ae4cd5-76d2-4e76-9528-6f511c4dca23" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.759045] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "refresh_cache-54ae4cd5-76d2-4e76-9528-6f511c4dca23" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 688.759191] env[65385]: DEBUG nova.network.neutron [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 688.822967] env[65385]: DEBUG nova.compute.utils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 688.827559] env[65385]: DEBUG nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 688.828083] env[65385]: DEBUG nova.network.neutron [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 688.828433] env[65385]: WARNING neutronclient.v2_0.client [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.828706] env[65385]: WARNING neutronclient.v2_0.client [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.829593] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.829954] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.844095] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c7adbc2-b559-4962-a610-cdcf9a06c30a tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.087s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.863374] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.863955] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.944390] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510508} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.947975] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 59f26752-75ab-45e5-b0a9-bbcef44ecd7f/59f26752-75ab-45e5-b0a9-bbcef44ecd7f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 688.948905] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 688.948905] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9eb72c08-dfa9-44c7-b43f-44eddc00a820 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.960165] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 688.960165] env[65385]: value = "task-4453342" [ 688.960165] env[65385]: _type = "Task" [ 688.960165] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.981071] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.003150] env[65385]: DEBUG nova.policy [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d5103a5769a4296973351194f49fed3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '754b1facaaa14501b2204c98e1d7a5a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 689.067052] env[65385]: DEBUG nova.network.neutron [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Updating instance_info_cache with network_info: [{"id": "29fa73f0-74dc-443a-b870-f4683fdc5d74", "address": "fa:16:3e:8c:f0:3d", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29fa73f0-74", "ovs_interfaceid": "29fa73f0-74dc-443a-b870-f4683fdc5d74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 689.125441] env[65385]: WARNING neutronclient.v2_0.client [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.126509] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.130021] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.266025] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.266025] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.332469] env[65385]: DEBUG nova.network.neutron [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updated VIF entry in instance network info cache for port 59289962-8634-414f-9bd2-b3f5b19af9b4. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 689.332865] env[65385]: DEBUG nova.network.neutron [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 689.338245] env[65385]: DEBUG nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 689.373804] env[65385]: DEBUG nova.network.neutron [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 689.473786] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.174085} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.474075] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 689.477215] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b657e31-1247-4dab-ad60-2d367ed1cac9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.500789] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 59f26752-75ab-45e5-b0a9-bbcef44ecd7f/59f26752-75ab-45e5-b0a9-bbcef44ecd7f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 689.502875] env[65385]: DEBUG nova.network.neutron [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Successfully created port: 9500dcc5-6478-4875-80fd-2ae8d2124100 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 689.506171] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.508810] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.512813] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38c050a1-ec1d-4103-bc2d-b333d2164ce8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.530198] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cffc00-3e3e-43f4-aaaa-932b05973c7e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.546350] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535c40fd-31cb-4195-a545-82d96d7e38ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.551205] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 689.551205] env[65385]: value = "task-4453343" [ 689.551205] env[65385]: _type = "Task" [ 689.551205] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.551484] env[65385]: DEBUG nova.network.neutron [-] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 689.588851] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "refresh_cache-0673a0d2-76ac-4cd5-8cc3-b8596877c641" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.590329] env[65385]: DEBUG nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Instance network_info: |[{"id": "29fa73f0-74dc-443a-b870-f4683fdc5d74", "address": "fa:16:3e:8c:f0:3d", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29fa73f0-74", "ovs_interfaceid": "29fa73f0-74dc-443a-b870-f4683fdc5d74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 689.596699] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:f0:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6815237d-f565-474d-a3c0-9c675478eb00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29fa73f0-74dc-443a-b870-f4683fdc5d74', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.606185] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 689.606300] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98c6267-5e2a-442e-8af8-c19b36a3d495 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.613086] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 689.613742] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453343, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.613967] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d3cc81f-1b30-4869-8696-4f571da45c78 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.630701] env[65385]: WARNING neutronclient.v2_0.client [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.634022] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.634022] env[65385]: WARNING openstack [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.644561] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73cf776-0faa-4b2c-be13-8fa9bb819f9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.650506] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.650506] env[65385]: value = "task-4453344" [ 689.650506] env[65385]: _type = "Task" [ 689.650506] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.663512] env[65385]: DEBUG nova.compute.provider_tree [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 689.673310] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453344, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.811213] env[65385]: DEBUG nova.network.neutron [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Updating instance_info_cache with network_info: [{"id": "6b67e7c9-05f5-47d9-9b81-86464d52033c", "address": "fa:16:3e:52:40:99", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b67e7c9-05", "ovs_interfaceid": "6b67e7c9-05f5-47d9-9b81-86464d52033c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 689.820564] env[65385]: DEBUG nova.compute.manager [req-c9663b9a-0fb9-4932-9b5a-8de5bcba8f8e req-3760f8db-d238-4303-b613-88ce33641af8 service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-vif-deleted-b89f2175-8529-42c4-9a78-6e7116ad7293 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 689.820564] env[65385]: DEBUG nova.compute.manager [req-c9663b9a-0fb9-4932-9b5a-8de5bcba8f8e req-3760f8db-d238-4303-b613-88ce33641af8 service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-vif-deleted-f23a0b59-d6a0-416e-a41c-d0f1f12672ee {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 689.820863] env[65385]: DEBUG nova.compute.manager [req-c9663b9a-0fb9-4932-9b5a-8de5bcba8f8e req-3760f8db-d238-4303-b613-88ce33641af8 service nova] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Received event network-vif-deleted-8c68958e-7ea9-4a06-b35a-05c48985738b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 689.838008] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.838343] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Received event network-vif-plugged-29fa73f0-74dc-443a-b870-f4683fdc5d74 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 689.838510] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Acquiring lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.838710] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.838862] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.839044] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] No waiting events found dispatching network-vif-plugged-29fa73f0-74dc-443a-b870-f4683fdc5d74 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 689.839199] env[65385]: WARNING nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Received unexpected event network-vif-plugged-29fa73f0-74dc-443a-b870-f4683fdc5d74 for instance with vm_state building and task_state spawning. [ 689.839354] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Received event network-changed-29fa73f0-74dc-443a-b870-f4683fdc5d74 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 689.839499] env[65385]: DEBUG nova.compute.manager [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Refreshing instance network info cache due to event network-changed-29fa73f0-74dc-443a-b870-f4683fdc5d74. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 689.839679] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Acquiring lock "refresh_cache-0673a0d2-76ac-4cd5-8cc3-b8596877c641" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.840126] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Acquired lock "refresh_cache-0673a0d2-76ac-4cd5-8cc3-b8596877c641" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.840126] env[65385]: DEBUG nova.network.neutron [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Refreshing network info cache for port 29fa73f0-74dc-443a-b870-f4683fdc5d74 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 690.063731] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453343, 'name': ReconfigVM_Task, 'duration_secs': 0.333322} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.064647] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 59f26752-75ab-45e5-b0a9-bbcef44ecd7f/59f26752-75ab-45e5-b0a9-bbcef44ecd7f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 690.064780] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-03c376dc-19c2-49ef-a57f-8cf8d4168733 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.072908] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 690.072908] env[65385]: value = "task-4453345" [ 690.072908] env[65385]: _type = "Task" [ 690.072908] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.082937] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453345, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.089143] env[65385]: INFO nova.compute.manager [-] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Took 1.62 seconds to deallocate network for instance. [ 690.164942] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453344, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.191392] env[65385]: ERROR nova.scheduler.client.report [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [req-aa17ee35-63dc-4089-b013-a4c3c1352048] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-aa17ee35-63dc-4089-b013-a4c3c1352048"}]} [ 690.215741] env[65385]: DEBUG nova.scheduler.client.report [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 690.235836] env[65385]: DEBUG nova.scheduler.client.report [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 690.235836] env[65385]: DEBUG nova.compute.provider_tree [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 690.250525] env[65385]: DEBUG nova.scheduler.client.report [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 690.272420] env[65385]: DEBUG nova.scheduler.client.report [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 690.317072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "refresh_cache-54ae4cd5-76d2-4e76-9528-6f511c4dca23" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 690.318327] env[65385]: DEBUG nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Instance network_info: |[{"id": "6b67e7c9-05f5-47d9-9b81-86464d52033c", "address": "fa:16:3e:52:40:99", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b67e7c9-05", "ovs_interfaceid": "6b67e7c9-05f5-47d9-9b81-86464d52033c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 690.319303] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:40:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b67e7c9-05f5-47d9-9b81-86464d52033c', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 690.333656] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating folder: Project (c27eb6232a2148a1a259f57494b4ae30). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.338343] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3edde452-a88b-4f94-bea7-2aaafc2a41d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.347356] env[65385]: WARNING neutronclient.v2_0.client [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.347356] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.348358] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.357037] env[65385]: DEBUG nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 690.359653] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created folder: Project (c27eb6232a2148a1a259f57494b4ae30) in parent group-v870881. [ 690.359760] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating folder: Instances. Parent ref: group-v870972. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.362839] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4aee3a5-aedc-4449-a41a-6dec93a27d05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.375431] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created folder: Instances in parent group-v870972. [ 690.375596] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 690.375692] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 690.376008] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9cfc0d76-cd29-4bfa-af20-3e2074a23876 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.401435] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 690.401435] env[65385]: value = "task-4453348" [ 690.401435] env[65385]: _type = "Task" [ 690.401435] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.403318] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 690.403470] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 690.403577] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 690.403762] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 690.403919] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 690.404080] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 690.404287] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 690.404438] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 690.404596] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 690.404772] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 690.404948] env[65385]: DEBUG nova.virt.hardware [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 690.405866] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a690b56-0842-4821-8c44-b281bc90b3e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.423764] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90c2680-77c1-47ac-a857-4108c9ce6da5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.427893] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453348, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.584746] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453345, 'name': Rename_Task, 'duration_secs': 0.156016} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.584937] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 690.585220] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6418a4b9-5986-4334-a0d2-e25c74034f27 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.594642] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 690.594642] env[65385]: value = "task-4453349" [ 690.594642] env[65385]: _type = "Task" [ 690.594642] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.601675] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.611386] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453349, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.667786] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453344, 'name': CreateVM_Task, 'duration_secs': 0.71981} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.667786] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 690.668422] env[65385]: WARNING neutronclient.v2_0.client [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.668868] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.669076] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.669842] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 690.670187] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cacee072-9c59-4f3f-a656-49bd307427dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.679800] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 690.679800] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263a57b-d87d-7adc-f541-612c84f7a313" [ 690.679800] env[65385]: _type = "Task" [ 690.679800] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.689852] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263a57b-d87d-7adc-f541-612c84f7a313, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.725060] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Acquiring lock "8ac44124-d8cf-421c-9ae4-c943df94550a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.725452] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.853679] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e371289-46b3-4093-8da1-73339903f1c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.867576] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968b7ccf-c7e3-4b06-820e-ff21f70e20ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.910021] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b02b6a-39ad-4203-82f3-da39a7d02b14 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.923524] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c1d399-e632-4bb9-aa0b-a8f997ea9086 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.932019] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453348, 'name': CreateVM_Task, 'duration_secs': 0.39292} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.932019] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 690.932019] env[65385]: WARNING neutronclient.v2_0.client [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.932019] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.942113] env[65385]: DEBUG nova.compute.provider_tree [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 691.106794] env[65385]: DEBUG oslo_vmware.api [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453349, 'name': PowerOnVM_Task, 'duration_secs': 0.496091} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.107206] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 691.107430] env[65385]: INFO nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Took 9.41 seconds to spawn the instance on the hypervisor. [ 691.107600] env[65385]: DEBUG nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 691.108441] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c3284c-9c6a-4ec1-80a7-15ed0aee629c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.128084] env[65385]: DEBUG nova.network.neutron [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Successfully updated port: 9500dcc5-6478-4875-80fd-2ae8d2124100 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 691.149955] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 691.150387] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 691.196559] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263a57b-d87d-7adc-f541-612c84f7a313, 'name': SearchDatastore_Task, 'duration_secs': 0.010566} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.196902] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.197395] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.197650] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.198766] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.198766] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 691.198766] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.198766] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 691.198924] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-46660b3f-59e2-4c66-bb44-883d2fc425a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.200889] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df443e2e-8279-41c6-97fd-bc375c2b255a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.207944] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 691.207944] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527d3e96-78d1-a492-01dd-3a9cf8ee9c34" [ 691.207944] env[65385]: _type = "Task" [ 691.207944] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.212466] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 691.212641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 691.213779] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e6f50da-b329-4e02-ab11-4cd2200652d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.219468] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527d3e96-78d1-a492-01dd-3a9cf8ee9c34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.223934] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 691.223934] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f10831-7d59-47b8-7879-98d6ed8f4705" [ 691.223934] env[65385]: _type = "Task" [ 691.223934] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.232223] env[65385]: INFO nova.compute.manager [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Detaching volume 3b3ed2bd-062e-4d87-8958-ddafcb01885d [ 691.245303] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f10831-7d59-47b8-7879-98d6ed8f4705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.291608] env[65385]: INFO nova.virt.block_device [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Attempting to driver detach volume 3b3ed2bd-062e-4d87-8958-ddafcb01885d from mountpoint /dev/sdb [ 691.292050] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 691.292127] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870959', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'name': 'volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ac44124-d8cf-421c-9ae4-c943df94550a', 'attached_at': '', 'detached_at': '', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'serial': '3b3ed2bd-062e-4d87-8958-ddafcb01885d'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 691.293399] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c92acf-d1b1-43ed-a03d-a5625513ba9e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.318981] env[65385]: WARNING neutronclient.v2_0.client [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 691.319966] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 691.321680] env[65385]: WARNING openstack [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 691.329824] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47d1ecc-a543-46e4-8c9c-62a29e275f64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.341684] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1941f6a2-83d7-47fc-b847-263ca9bda6a0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.366151] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffd62cf-7f04-4451-a739-74e201088609 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.383569] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] The volume has not been displaced from its original location: [datastore1] volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d/volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 691.389384] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Reconfiguring VM instance instance-0000000a to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 691.389883] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2375cbdf-c652-4c22-aa2c-767261c769e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.409214] env[65385]: DEBUG oslo_vmware.api [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Waiting for the task: (returnval){ [ 691.409214] env[65385]: value = "task-4453350" [ 691.409214] env[65385]: _type = "Task" [ 691.409214] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.422324] env[65385]: DEBUG oslo_vmware.api [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453350, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.439122] env[65385]: DEBUG nova.compute.manager [req-9f77e0f7-75d2-4a20-9129-f86d78dede4a req-a44a0e61-1ab3-40cc-ad12-3d73e773fcde service nova] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Received event network-vif-deleted-07665fa0-2bf4-4c1d-9058-505e323caa56 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 691.488927] env[65385]: DEBUG nova.network.neutron [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Updated VIF entry in instance network info cache for port 29fa73f0-74dc-443a-b870-f4683fdc5d74. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 691.489308] env[65385]: DEBUG nova.network.neutron [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Updating instance_info_cache with network_info: [{"id": "29fa73f0-74dc-443a-b870-f4683fdc5d74", "address": "fa:16:3e:8c:f0:3d", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29fa73f0-74", "ovs_interfaceid": "29fa73f0-74dc-443a-b870-f4683fdc5d74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 691.491418] env[65385]: DEBUG nova.scheduler.client.report [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 50 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 691.491660] env[65385]: DEBUG nova.compute.provider_tree [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 50 to 51 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 691.491922] env[65385]: DEBUG nova.compute.provider_tree [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 691.634196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-28fd3828-38ec-426f-b4dc-1e094304780b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.634196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-28fd3828-38ec-426f-b4dc-1e094304780b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.634196] env[65385]: DEBUG nova.network.neutron [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 691.640497] env[65385]: INFO nova.compute.manager [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Took 26.84 seconds to build instance. [ 691.720565] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527d3e96-78d1-a492-01dd-3a9cf8ee9c34, 'name': SearchDatastore_Task, 'duration_secs': 0.013568} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.720916] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.721192] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.721766] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.733913] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f10831-7d59-47b8-7879-98d6ed8f4705, 'name': SearchDatastore_Task, 'duration_secs': 0.029365} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.735092] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6342f32-4514-4bff-bf8c-24f499a87003 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.741118] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 691.741118] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520d89f7-522f-86a5-1c2f-e3f9c20cce0e" [ 691.741118] env[65385]: _type = "Task" [ 691.741118] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.749754] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520d89f7-522f-86a5-1c2f-e3f9c20cce0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.920214] env[65385]: DEBUG oslo_vmware.api [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453350, 'name': ReconfigVM_Task, 'duration_secs': 0.278723} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.920800] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Reconfigured VM instance instance-0000000a to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 691.925630] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89fc93a5-cf95-47d3-adb8-407b07a3c636 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.942740] env[65385]: DEBUG oslo_vmware.api [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Waiting for the task: (returnval){ [ 691.942740] env[65385]: value = "task-4453351" [ 691.942740] env[65385]: _type = "Task" [ 691.942740] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.952141] env[65385]: DEBUG oslo_vmware.api [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.997831] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ad7cce9-526d-4607-b325-934a53ad070a req-ee7a29e0-9402-4246-8a4c-5c9da7e72292 service nova] Releasing lock "refresh_cache-0673a0d2-76ac-4cd5-8cc3-b8596877c641" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.033407] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "60067529-1071-4295-b1c9-21523bf347d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.033560] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "60067529-1071-4295-b1c9-21523bf347d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.034306] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "60067529-1071-4295-b1c9-21523bf347d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.034306] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "60067529-1071-4295-b1c9-21523bf347d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.034306] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "60067529-1071-4295-b1c9-21523bf347d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.037388] env[65385]: INFO nova.compute.manager [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Terminating instance [ 692.138605] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.139168] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.147496] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00ce743d-3160-4292-9f81-e3a4061a4c0e tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.356s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.218661] env[65385]: DEBUG nova.network.neutron [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 692.251528] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.251911] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.268339] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520d89f7-522f-86a5-1c2f-e3f9c20cce0e, 'name': SearchDatastore_Task, 'duration_secs': 0.018998} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.268653] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.269156] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0673a0d2-76ac-4cd5-8cc3-b8596877c641/0673a0d2-76ac-4cd5-8cc3-b8596877c641.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 692.269387] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.269517] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 692.269768] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88bd1c96-a2d6-4eaa-9dda-4b1e48e857bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.274145] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66fcd7df-8ff5-4adc-9d46-0db62cd812db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.286038] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 692.286038] env[65385]: value = "task-4453352" [ 692.286038] env[65385]: _type = "Task" [ 692.286038] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.286927] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 692.287346] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 692.293271] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fa027ba-fd2c-4868-8fb6-64de39b99003 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.302035] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 692.302035] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ad1cad-b525-bc65-df56-3720cbd09cc8" [ 692.302035] env[65385]: _type = "Task" [ 692.302035] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.308492] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453352, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.321321] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ad1cad-b525-bc65-df56-3720cbd09cc8, 'name': SearchDatastore_Task, 'duration_secs': 0.013479} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.321418] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0e9af71-d8ac-4a47-b650-4ebdd9df355b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.329861] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 692.329861] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5237ee8f-19c1-cd6e-9d9f-17b95ba98037" [ 692.329861] env[65385]: _type = "Task" [ 692.329861] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.341261] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5237ee8f-19c1-cd6e-9d9f-17b95ba98037, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.378054] env[65385]: DEBUG nova.objects.instance [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lazy-loading 'flavor' on Instance uuid 649b4e21-fc55-415c-a1f3-ec724397b874 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 692.416251] env[65385]: WARNING neutronclient.v2_0.client [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.416941] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.417367] env[65385]: WARNING openstack [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.460029] env[65385]: DEBUG oslo_vmware.api [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Task: {'id': task-4453351, 'name': ReconfigVM_Task, 'duration_secs': 0.157398} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.460029] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870959', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'name': 'volume-3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8ac44124-d8cf-421c-9ae4-c943df94550a', 'attached_at': '', 'detached_at': '', 'volume_id': '3b3ed2bd-062e-4d87-8958-ddafcb01885d', 'serial': '3b3ed2bd-062e-4d87-8958-ddafcb01885d'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 692.507072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 4.186s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.510402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.293s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.510767] env[65385]: DEBUG nova.objects.instance [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lazy-loading 'resources' on Instance uuid a8c76329-5056-4cf0-ac32-85d46429d3db {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 692.545790] env[65385]: DEBUG nova.compute.manager [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 692.545790] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 692.546553] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815d1680-f100-41e3-aada-4c3a50757e8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.556694] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 692.556863] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c0ff476-4d21-4f1e-ae7a-0a004e78e0d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.570466] env[65385]: DEBUG oslo_vmware.api [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 692.570466] env[65385]: value = "task-4453353" [ 692.570466] env[65385]: _type = "Task" [ 692.570466] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.578854] env[65385]: DEBUG nova.network.neutron [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Updating instance_info_cache with network_info: [{"id": "9500dcc5-6478-4875-80fd-2ae8d2124100", "address": "fa:16:3e:e9:7c:72", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9500dcc5-64", "ovs_interfaceid": "9500dcc5-6478-4875-80fd-2ae8d2124100", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 692.589763] env[65385]: DEBUG oslo_vmware.api [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.801671] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453352, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.844903] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5237ee8f-19c1-cd6e-9d9f-17b95ba98037, 'name': SearchDatastore_Task, 'duration_secs': 0.012814} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.845346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.845647] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 54ae4cd5-76d2-4e76-9528-6f511c4dca23/54ae4cd5-76d2-4e76-9528-6f511c4dca23.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 692.845956] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-748cc370-2142-4660-a2b6-79a1d4ee9dc1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.854766] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 692.854766] env[65385]: value = "task-4453354" [ 692.854766] env[65385]: _type = "Task" [ 692.854766] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.869419] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.888782] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.888964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquired lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.889171] env[65385]: DEBUG nova.network.neutron [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 692.889358] env[65385]: DEBUG nova.objects.instance [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lazy-loading 'info_cache' on Instance uuid 649b4e21-fc55-415c-a1f3-ec724397b874 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 692.917494] env[65385]: DEBUG nova.compute.manager [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Received event network-vif-plugged-6b67e7c9-05f5-47d9-9b81-86464d52033c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 692.917924] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Acquiring lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.919076] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.919419] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.919726] env[65385]: DEBUG nova.compute.manager [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] No waiting events found dispatching network-vif-plugged-6b67e7c9-05f5-47d9-9b81-86464d52033c {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 692.920050] env[65385]: WARNING nova.compute.manager [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Received unexpected event network-vif-plugged-6b67e7c9-05f5-47d9-9b81-86464d52033c for instance with vm_state building and task_state spawning. [ 692.920255] env[65385]: DEBUG nova.compute.manager [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Received event network-changed-6b67e7c9-05f5-47d9-9b81-86464d52033c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 692.920485] env[65385]: DEBUG nova.compute.manager [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Refreshing instance network info cache due to event network-changed-6b67e7c9-05f5-47d9-9b81-86464d52033c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 692.920642] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Acquiring lock "refresh_cache-54ae4cd5-76d2-4e76-9528-6f511c4dca23" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.920860] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Acquired lock "refresh_cache-54ae4cd5-76d2-4e76-9528-6f511c4dca23" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.921155] env[65385]: DEBUG nova.network.neutron [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Refreshing network info cache for port 6b67e7c9-05f5-47d9-9b81-86464d52033c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 693.058708] env[65385]: DEBUG nova.objects.instance [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lazy-loading 'flavor' on Instance uuid 8ac44124-d8cf-421c-9ae4-c943df94550a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 693.082784] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-28fd3828-38ec-426f-b4dc-1e094304780b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 693.082784] env[65385]: DEBUG nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Instance network_info: |[{"id": "9500dcc5-6478-4875-80fd-2ae8d2124100", "address": "fa:16:3e:e9:7c:72", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9500dcc5-64", "ovs_interfaceid": "9500dcc5-6478-4875-80fd-2ae8d2124100", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 693.087211] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:7c:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9500dcc5-6478-4875-80fd-2ae8d2124100', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 693.094636] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 693.094952] env[65385]: DEBUG oslo_vmware.api [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453353, 'name': PowerOffVM_Task, 'duration_secs': 0.395975} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.095592] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 693.096094] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 693.096094] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 693.096321] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d796839c-9adb-41a6-b385-f9d2d25c0c8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.112276] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5184a4e4-b356-4821-8b07-3dfff2db90d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.117624] env[65385]: INFO nova.scheduler.client.report [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleted allocation for migration cbca16ba-15ec-46bc-8a77-271b74b9f582 [ 693.130333] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 693.130333] env[65385]: value = "task-4453359" [ 693.130333] env[65385]: _type = "Task" [ 693.130333] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.152274] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453359, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.232791] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 693.233200] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 693.233383] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleting the datastore file [datastore2] 60067529-1071-4295-b1c9-21523bf347d9 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 693.233765] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20fa01d8-84aa-4001-9777-b1d358ff4822 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.248508] env[65385]: DEBUG oslo_vmware.api [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 693.248508] env[65385]: value = "task-4453360" [ 693.248508] env[65385]: _type = "Task" [ 693.248508] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.262065] env[65385]: DEBUG oslo_vmware.api [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453360, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.303617] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453352, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568395} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.303617] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0673a0d2-76ac-4cd5-8cc3-b8596877c641/0673a0d2-76ac-4cd5-8cc3-b8596877c641.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 693.303617] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.303913] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2a75d2e-e66b-4b4c-a937-5c20d6dc5d5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.316409] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 693.316409] env[65385]: value = "task-4453361" [ 693.316409] env[65385]: _type = "Task" [ 693.316409] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.329697] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453361, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.369592] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453354, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.399337] env[65385]: DEBUG nova.objects.base [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Object Instance<649b4e21-fc55-415c-a1f3-ec724397b874> lazy-loaded attributes: flavor,info_cache {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 693.429683] env[65385]: WARNING neutronclient.v2_0.client [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.430514] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.430871] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.531876] env[65385]: INFO nova.compute.manager [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Rebuilding instance [ 693.588196] env[65385]: DEBUG nova.compute.manager [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 693.589414] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df743a2-854c-4dee-840f-6ff4aadd402e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.624905] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4541c10-9684-4381-933b-c03d90b47da5 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 27.235s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 693.643355] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453359, 'name': CreateVM_Task, 'duration_secs': 0.426624} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.645889] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 693.646516] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55011a27-446b-4807-b841-92dea4db088b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.650267] env[65385]: WARNING neutronclient.v2_0.client [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.651401] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.651613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 693.652027] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 693.652582] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c603597-24dd-41c5-aed3-d9578e910f85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.661833] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c274bd-20d0-4c6f-ae1f-4688993ce319 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.665332] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 693.665332] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b03cd9-a84a-5162-d55e-bf1209b7e2ca" [ 693.665332] env[65385]: _type = "Task" [ 693.665332] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.696968] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0009db93-0c3c-43ea-b858-ff647bf9c244 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.703411] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b03cd9-a84a-5162-d55e-bf1209b7e2ca, 'name': SearchDatastore_Task, 'duration_secs': 0.033819} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.704173] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 693.704422] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 693.704839] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.704839] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 693.704965] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 693.705249] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddefb00e-20ce-4302-a5e6-3fdf48eeb063 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.710882] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae73c11a-fd13-4b03-bde5-fc3f7792bbbf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.718603] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 693.718603] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 693.727424] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fa0d8bf-a5d2-4cce-88ed-8a41e561eb55 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.730010] env[65385]: DEBUG nova.compute.provider_tree [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.736157] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 693.736157] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52048789-bf89-22fd-da0d-d66fe2db244f" [ 693.736157] env[65385]: _type = "Task" [ 693.736157] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.746281] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52048789-bf89-22fd-da0d-d66fe2db244f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.759539] env[65385]: DEBUG oslo_vmware.api [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453360, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.828027] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453361, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099784} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.828396] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 693.829214] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6071943-3387-4b27-9de7-1221078f5885 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.854424] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 0673a0d2-76ac-4cd5-8cc3-b8596877c641/0673a0d2-76ac-4cd5-8cc3-b8596877c641.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 693.854889] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bebb6d29-6aa8-4c6c-99ee-a7db876ece2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.882213] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62834} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.883999] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 54ae4cd5-76d2-4e76-9528-6f511c4dca23/54ae4cd5-76d2-4e76-9528-6f511c4dca23.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 693.884244] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.884568] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 693.884568] env[65385]: value = "task-4453362" [ 693.884568] env[65385]: _type = "Task" [ 693.884568] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.884828] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a33bd1e7-ddec-4be1-b88e-65791ee077dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.901340] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453362, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.903286] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 693.903286] env[65385]: value = "task-4453363" [ 693.903286] env[65385]: _type = "Task" [ 693.903286] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.907293] env[65385]: WARNING neutronclient.v2_0.client [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.908063] env[65385]: WARNING openstack [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.908464] env[65385]: WARNING openstack [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.921919] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453363, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.067385] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dbc0887c-d399-471d-8248-b8be4b6c74c5 tempest-VolumesAssistedSnapshotsTest-1920711293 tempest-VolumesAssistedSnapshotsTest-1920711293-project-admin] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.342s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.133393] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.133777] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.233570] env[65385]: DEBUG nova.scheduler.client.report [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 694.260194] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52048789-bf89-22fd-da0d-d66fe2db244f, 'name': SearchDatastore_Task, 'duration_secs': 0.018965} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.262903] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17cd7c1b-beac-44e6-8b72-71564abda358 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.270095] env[65385]: DEBUG oslo_vmware.api [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453360, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.646533} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.270884] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 694.271204] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 694.271305] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.271799] env[65385]: INFO nova.compute.manager [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Took 1.73 seconds to destroy the instance on the hypervisor. [ 694.271799] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 694.271974] env[65385]: DEBUG nova.compute.manager [-] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 694.272297] env[65385]: DEBUG nova.network.neutron [-] [instance: 60067529-1071-4295-b1c9-21523bf347d9] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 694.272375] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.273020] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.273326] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.283924] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 694.283924] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521d8dca-6604-7eb7-75c9-da173af2e52b" [ 694.283924] env[65385]: _type = "Task" [ 694.283924] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.295295] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521d8dca-6604-7eb7-75c9-da173af2e52b, 'name': SearchDatastore_Task, 'duration_secs': 0.015585} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.295555] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 694.295879] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 28fd3828-38ec-426f-b4dc-1e094304780b/28fd3828-38ec-426f-b4dc-1e094304780b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 694.296557] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b602648f-f398-4661-b2c7-516aebbc3ab8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.304617] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 694.304617] env[65385]: value = "task-4453365" [ 694.304617] env[65385]: _type = "Task" [ 694.304617] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.315518] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.346142] env[65385]: WARNING openstack [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.346886] env[65385]: WARNING openstack [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.360692] env[65385]: WARNING neutronclient.v2_0.client [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.361376] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.361783] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.400181] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453362, 'name': ReconfigVM_Task, 'duration_secs': 0.323587} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.403537] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 0673a0d2-76ac-4cd5-8cc3-b8596877c641/0673a0d2-76ac-4cd5-8cc3-b8596877c641.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 694.403537] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00c33563-0d40-4a9d-8856-f64e54d73cda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.409180] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.414562] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 694.414562] env[65385]: value = "task-4453366" [ 694.414562] env[65385]: _type = "Task" [ 694.414562] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.420843] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453363, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103586} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.426285] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 694.427352] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5a76e5-b0d1-4858-b65c-cede7066dc61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.439015] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453366, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.459920] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 54ae4cd5-76d2-4e76-9528-6f511c4dca23/54ae4cd5-76d2-4e76-9528-6f511c4dca23.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 694.460455] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0f307cf-8acc-4b6e-9c4e-cde6418a84a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.488044] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 694.488044] env[65385]: value = "task-4453367" [ 694.488044] env[65385]: _type = "Task" [ 694.488044] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.498837] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.607023] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 694.607406] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61b7aabe-374b-4084-b0f7-3b2b4d55d464 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.615854] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 694.615854] env[65385]: value = "task-4453368" [ 694.615854] env[65385]: _type = "Task" [ 694.615854] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.627187] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453368, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.750592] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.240s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.753681] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.188s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.755405] env[65385]: INFO nova.compute.claims [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.786385] env[65385]: INFO nova.scheduler.client.report [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Deleted allocations for instance a8c76329-5056-4cf0-ac32-85d46429d3db [ 694.821967] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453365, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.914506] env[65385]: WARNING neutronclient.v2_0.client [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.915194] env[65385]: WARNING openstack [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.916099] env[65385]: WARNING openstack [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.938873] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453366, 'name': Rename_Task, 'duration_secs': 0.172527} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.939254] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 694.939555] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90b8ad07-f636-4ecf-ae30-487da0bcd4c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.950123] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 694.950123] env[65385]: value = "task-4453369" [ 694.950123] env[65385]: _type = "Task" [ 694.950123] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.963461] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.990434] env[65385]: DEBUG nova.network.neutron [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Updated VIF entry in instance network info cache for port 6b67e7c9-05f5-47d9-9b81-86464d52033c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 694.990434] env[65385]: DEBUG nova.network.neutron [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Updating instance_info_cache with network_info: [{"id": "6b67e7c9-05f5-47d9-9b81-86464d52033c", "address": "fa:16:3e:52:40:99", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b67e7c9-05", "ovs_interfaceid": "6b67e7c9-05f5-47d9-9b81-86464d52033c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.007896] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453367, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.126572] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453368, 'name': PowerOffVM_Task, 'duration_secs': 0.117322} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.126977] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 695.128778] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 695.129665] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d387863-2da7-45b2-8ed7-f502bdfdeec4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.137977] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 695.138418] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87f12299-89ce-4352-b7ca-1819c95268c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.169175] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 695.169175] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 695.169175] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Deleting the datastore file [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 695.169175] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e7519d0-6737-45e1-9515-decf01244c78 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.177618] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 695.177618] env[65385]: value = "task-4453371" [ 695.177618] env[65385]: _type = "Task" [ 695.177618] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.188542] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.214328] env[65385]: DEBUG nova.network.neutron [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Updating instance_info_cache with network_info: [{"id": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "address": "fa:16:3e:5c:e2:96", "network": {"id": "22005ece-dcae-4f26-bb66-040863ebbf8f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1035018373-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "112402342b0e481f83be7e17c9f739e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a4f4b0b-f7", "ovs_interfaceid": "7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.295957] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7f9cfb16-0069-4c44-9889-95a0827f6738 tempest-ServersTestManualDisk-1709045078 tempest-ServersTestManualDisk-1709045078-project-member] Lock "a8c76329-5056-4cf0-ac32-85d46429d3db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.928s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.320950] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.754995} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.321451] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 28fd3828-38ec-426f-b4dc-1e094304780b/28fd3828-38ec-426f-b4dc-1e094304780b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 695.321718] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 695.322028] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-677ceb37-fe96-4d38-975f-3690c0dc996d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.330850] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 695.330850] env[65385]: value = "task-4453372" [ 695.330850] env[65385]: _type = "Task" [ 695.330850] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.341169] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453372, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.347533] env[65385]: DEBUG nova.network.neutron [-] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.432980] env[65385]: DEBUG nova.compute.manager [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Received event network-vif-plugged-9500dcc5-6478-4875-80fd-2ae8d2124100 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 695.433572] env[65385]: DEBUG oslo_concurrency.lockutils [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Acquiring lock "28fd3828-38ec-426f-b4dc-1e094304780b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.433729] env[65385]: DEBUG oslo_concurrency.lockutils [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Lock "28fd3828-38ec-426f-b4dc-1e094304780b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.433948] env[65385]: DEBUG oslo_concurrency.lockutils [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Lock "28fd3828-38ec-426f-b4dc-1e094304780b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.434267] env[65385]: DEBUG nova.compute.manager [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] No waiting events found dispatching network-vif-plugged-9500dcc5-6478-4875-80fd-2ae8d2124100 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 695.434561] env[65385]: WARNING nova.compute.manager [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Received unexpected event network-vif-plugged-9500dcc5-6478-4875-80fd-2ae8d2124100 for instance with vm_state building and task_state spawning. [ 695.434742] env[65385]: DEBUG nova.compute.manager [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Received event network-changed-9500dcc5-6478-4875-80fd-2ae8d2124100 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 695.434989] env[65385]: DEBUG nova.compute.manager [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Refreshing instance network info cache due to event network-changed-9500dcc5-6478-4875-80fd-2ae8d2124100. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 695.435282] env[65385]: DEBUG oslo_concurrency.lockutils [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Acquiring lock "refresh_cache-28fd3828-38ec-426f-b4dc-1e094304780b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.435485] env[65385]: DEBUG oslo_concurrency.lockutils [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Acquired lock "refresh_cache-28fd3828-38ec-426f-b4dc-1e094304780b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.435910] env[65385]: DEBUG nova.network.neutron [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Refreshing network info cache for port 9500dcc5-6478-4875-80fd-2ae8d2124100 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 695.465848] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453369, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.498348] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Releasing lock "refresh_cache-54ae4cd5-76d2-4e76-9528-6f511c4dca23" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.498740] env[65385]: DEBUG nova.compute.manager [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Received event network-changed-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 695.498889] env[65385]: DEBUG nova.compute.manager [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Refreshing instance network info cache due to event network-changed-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 695.499187] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Acquiring lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.499261] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Acquired lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.499444] env[65385]: DEBUG nova.network.neutron [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Refreshing network info cache for port d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 695.510680] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453367, 'name': ReconfigVM_Task, 'duration_secs': 0.878554} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.511916] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 54ae4cd5-76d2-4e76-9528-6f511c4dca23/54ae4cd5-76d2-4e76-9528-6f511c4dca23.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 695.514201] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66f25b75-88f6-446a-9e81-3e29706eab73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.524228] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 695.524228] env[65385]: value = "task-4453373" [ 695.524228] env[65385]: _type = "Task" [ 695.524228] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.543384] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453373, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.688916] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237997} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.688916] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 695.689116] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 695.689291] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 695.719914] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Releasing lock "refresh_cache-649b4e21-fc55-415c-a1f3-ec724397b874" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.844957] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453372, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105065} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.845317] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 695.846271] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412f682e-3e8d-4fa2-aedf-d89eff0fe955 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.851990] env[65385]: INFO nova.compute.manager [-] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Took 1.58 seconds to deallocate network for instance. [ 695.878780] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 28fd3828-38ec-426f-b4dc-1e094304780b/28fd3828-38ec-426f-b4dc-1e094304780b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.882384] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b9a323d-af52-4b0a-92fd-847380cbebae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.907605] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 695.907605] env[65385]: value = "task-4453374" [ 695.907605] env[65385]: _type = "Task" [ 695.907605] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.921626] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453374, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.940720] env[65385]: WARNING neutronclient.v2_0.client [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 695.943073] env[65385]: WARNING openstack [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.943073] env[65385]: WARNING openstack [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.968695] env[65385]: DEBUG oslo_vmware.api [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453369, 'name': PowerOnVM_Task, 'duration_secs': 0.531996} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.968695] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 695.968695] env[65385]: INFO nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Took 11.00 seconds to spawn the instance on the hypervisor. [ 695.968695] env[65385]: DEBUG nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 695.968695] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297cf9bc-6a88-4016-88b3-8b43b30eb8de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.012511] env[65385]: WARNING neutronclient.v2_0.client [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.013271] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.013647] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.037340] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453373, 'name': Rename_Task, 'duration_secs': 0.398588} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.037718] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 696.041727] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65920ac5-bd72-4c4d-b0b1-f376102039f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.052663] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 696.052663] env[65385]: value = "task-4453375" [ 696.052663] env[65385]: _type = "Task" [ 696.052663] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.070488] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453375, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.228694] env[65385]: WARNING openstack [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.228694] env[65385]: WARNING openstack [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.350452] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.350829] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.362032] env[65385]: WARNING neutronclient.v2_0.client [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.362032] env[65385]: WARNING openstack [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.362032] env[65385]: WARNING openstack [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.373665] env[65385]: DEBUG nova.compute.manager [req-b474f529-6b93-4f0b-9c1c-6aef164d3a8c req-f12e14fb-f3b1-49ca-8e64-8e819e4000c2 service nova] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Received event network-vif-deleted-82373c5b-5d69-406c-9fca-ed283c2a5f88 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 696.382451] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.395152] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b3f9ae-100e-4866-8855-577f11f5c1fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.409229] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa2d407-fe2f-4130-988a-2badbae95821 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.430948] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453374, 'name': ReconfigVM_Task, 'duration_secs': 0.386556} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.460241] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 28fd3828-38ec-426f-b4dc-1e094304780b/28fd3828-38ec-426f-b4dc-1e094304780b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 696.465779] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8261ae5-8e48-4709-b93b-5b87aaa33df4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.469019] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8f3211-040f-4936-aadd-8f093564a77b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.471945] env[65385]: WARNING neutronclient.v2_0.client [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.472612] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.473340] env[65385]: WARNING openstack [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.501212] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 696.501212] env[65385]: value = "task-4453377" [ 696.501212] env[65385]: _type = "Task" [ 696.501212] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.502278] env[65385]: INFO nova.compute.manager [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Took 29.79 seconds to build instance. [ 696.505348] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9958f7-250c-4a9a-ad32-50a28a8041b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.527542] env[65385]: DEBUG nova.compute.provider_tree [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 696.533874] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453377, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.540742] env[65385]: DEBUG nova.network.neutron [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Updated VIF entry in instance network info cache for port 9500dcc5-6478-4875-80fd-2ae8d2124100. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 696.541153] env[65385]: DEBUG nova.network.neutron [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Updating instance_info_cache with network_info: [{"id": "9500dcc5-6478-4875-80fd-2ae8d2124100", "address": "fa:16:3e:e9:7c:72", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9500dcc5-64", "ovs_interfaceid": "9500dcc5-6478-4875-80fd-2ae8d2124100", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.567175] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453375, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.632869] env[65385]: DEBUG nova.network.neutron [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Updated VIF entry in instance network info cache for port d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 696.633675] env[65385]: DEBUG nova.network.neutron [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Updating instance_info_cache with network_info: [{"id": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "address": "fa:16:3e:9d:e2:9e", "network": {"id": "a97d30fb-abe7-4ead-a564-a64f533b6756", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1681302515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c8167324e3664a578a5ca8c8a04c1241", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "537e0890-4fa2-4f2d-b74c-49933a4edf53", "external-id": "nsx-vlan-transportzone-82", "segmentation_id": 82, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd970616a-b8", "ovs_interfaceid": "d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.737601] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 696.737873] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 696.738034] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 696.738210] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 696.738344] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 696.738478] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 696.738673] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.738873] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 696.738983] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 696.739147] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 696.739307] env[65385]: DEBUG nova.virt.hardware [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 696.739980] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 696.740845] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7e9dc0-fa78-47f3-9a20-cbc17fc2ac19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.744070] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01f96e92-2f66-4918-81c1-69b2cdda43e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.755752] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57287777-2cd7-483b-94de-0c4c1012abaa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.762047] env[65385]: DEBUG oslo_vmware.api [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 696.762047] env[65385]: value = "task-4453378" [ 696.762047] env[65385]: _type = "Task" [ 696.762047] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.775489] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 696.781680] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 696.782538] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 696.782790] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b67b4834-936d-4652-bb59-b324c23c53b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.800258] env[65385]: DEBUG oslo_vmware.api [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453378, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.807103] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 696.807103] env[65385]: value = "task-4453379" [ 696.807103] env[65385]: _type = "Task" [ 696.807103] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.818211] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453379, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.988268] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "c798006c-3e74-461f-8397-3d8ca0be0a95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.988713] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.012689] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8903e91d-61b7-49de-99ce-7db1279c2af3 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.324s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.018409] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453377, 'name': Rename_Task, 'duration_secs': 0.176314} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.018726] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 697.019069] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86cd3f8f-8d54-4a80-aca4-366ba5348765 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.027484] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 697.027484] env[65385]: value = "task-4453380" [ 697.027484] env[65385]: _type = "Task" [ 697.027484] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.039980] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.044053] env[65385]: DEBUG oslo_concurrency.lockutils [req-728c32b7-4074-4df3-a5d3-72a628601e86 req-dba3bedc-26de-476b-8c77-83611eb59223 service nova] Releasing lock "refresh_cache-28fd3828-38ec-426f-b4dc-1e094304780b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 697.064620] env[65385]: ERROR nova.scheduler.client.report [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [req-750b4624-bd34-4b0f-a465-72ff5f768579] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-750b4624-bd34-4b0f-a465-72ff5f768579"}]} [ 697.074355] env[65385]: DEBUG oslo_vmware.api [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453375, 'name': PowerOnVM_Task, 'duration_secs': 0.957293} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.074625] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 697.074884] env[65385]: INFO nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Took 9.27 seconds to spawn the instance on the hypervisor. [ 697.075089] env[65385]: DEBUG nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 697.075934] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35571d57-8793-4ae8-bf27-7bf171e76402 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.095019] env[65385]: DEBUG nova.scheduler.client.report [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 697.112302] env[65385]: DEBUG nova.scheduler.client.report [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 697.112513] env[65385]: DEBUG nova.compute.provider_tree [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 697.126222] env[65385]: DEBUG nova.scheduler.client.report [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 697.136317] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a113112-86a9-454e-8afa-0ba6e7135521 req-6af3f3a6-1c2b-464b-89bc-c59e58412c0a service nova] Releasing lock "refresh_cache-2e4c064f-27b0-4ccb-8fe6-c34a61153663" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 697.147087] env[65385]: DEBUG nova.scheduler.client.report [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 697.280708] env[65385]: DEBUG oslo_vmware.api [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453378, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.326086] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453379, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.492041] env[65385]: DEBUG nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 697.543162] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453380, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.601491] env[65385]: INFO nova.compute.manager [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Took 30.13 seconds to build instance. [ 697.705594] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1c06d4-db14-4ff4-ac8d-9fb60f1abea7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.713915] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dd98a9-0ffb-4ef0-81c6-adf427bcec86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.747698] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af630a4-c995-43e2-bedd-26188d2f52be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.756472] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b792c346-a8db-44c3-adcb-453e1f224d9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.771759] env[65385]: DEBUG nova.compute.provider_tree [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 697.781811] env[65385]: DEBUG oslo_vmware.api [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453378, 'name': PowerOnVM_Task, 'duration_secs': 0.602157} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.782735] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 697.782903] env[65385]: DEBUG nova.compute.manager [None req-c85e34c7-a9eb-42f4-95aa-373093a72162 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 697.783665] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492b1686-959b-4b29-9712-91b944a1d0d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.822590] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453379, 'name': CreateVM_Task, 'duration_secs': 0.602283} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.822861] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 697.823231] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.823388] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.824057] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 697.824057] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7d97713-c903-415b-a27f-02bc7538dd53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.830415] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 697.830415] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521d441e-6ae1-0695-bc23-2f143106283d" [ 697.830415] env[65385]: _type = "Task" [ 697.830415] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.840818] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521d441e-6ae1-0695-bc23-2f143106283d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.026697] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.041947] env[65385]: DEBUG oslo_vmware.api [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453380, 'name': PowerOnVM_Task, 'duration_secs': 0.545438} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.042556] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 698.042824] env[65385]: INFO nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Took 7.69 seconds to spawn the instance on the hypervisor. [ 698.043133] env[65385]: DEBUG nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 698.044394] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1f8c3c-ce72-42bb-88f5-f256a062489a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.095246] env[65385]: DEBUG nova.compute.manager [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Received event network-changed-59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 698.095246] env[65385]: DEBUG nova.compute.manager [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Refreshing instance network info cache due to event network-changed-59289962-8634-414f-9bd2-b3f5b19af9b4. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 698.095246] env[65385]: DEBUG oslo_concurrency.lockutils [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.095246] env[65385]: DEBUG oslo_concurrency.lockutils [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.095847] env[65385]: DEBUG nova.network.neutron [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Refreshing network info cache for port 59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 698.107215] env[65385]: DEBUG oslo_concurrency.lockutils [None req-228182bd-44c5-4dd4-9d6d-a446ccc9195a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.649s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.337377] env[65385]: DEBUG nova.scheduler.client.report [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 53 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 698.337707] env[65385]: DEBUG nova.compute.provider_tree [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 53 to 54 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 698.337941] env[65385]: DEBUG nova.compute.provider_tree [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 698.350228] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521d441e-6ae1-0695-bc23-2f143106283d, 'name': SearchDatastore_Task, 'duration_secs': 0.036673} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.350965] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.350965] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.351410] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.351633] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.351902] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.353533] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8893fefb-66c4-4fd6-b33b-63b2bbcf2334 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.364642] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.364870] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 698.366026] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e622743-3c57-4832-a253-427d276cc0f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.373164] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 698.373164] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d0c13b-01bc-c952-19c4-8ef149dba1f4" [ 698.373164] env[65385]: _type = "Task" [ 698.373164] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.383536] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d0c13b-01bc-c952-19c4-8ef149dba1f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.577342] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "8ac44124-d8cf-421c-9ae4-c943df94550a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.577342] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.578539] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "8ac44124-d8cf-421c-9ae4-c943df94550a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.578539] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.578539] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.581849] env[65385]: INFO nova.compute.manager [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Took 29.67 seconds to build instance. [ 698.583138] env[65385]: INFO nova.compute.manager [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Terminating instance [ 698.603415] env[65385]: WARNING neutronclient.v2_0.client [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.604129] env[65385]: WARNING openstack [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.604855] env[65385]: WARNING openstack [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.808602] env[65385]: WARNING openstack [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.809056] env[65385]: WARNING openstack [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.847707] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.094s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.848253] env[65385]: DEBUG nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 698.852489] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.926s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.852779] env[65385]: DEBUG nova.objects.instance [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lazy-loading 'resources' on Instance uuid bf79fbcd-e943-4ff3-bcf8-6a2817df8292 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 698.890589] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d0c13b-01bc-c952-19c4-8ef149dba1f4, 'name': SearchDatastore_Task, 'duration_secs': 0.013646} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.896741] env[65385]: WARNING neutronclient.v2_0.client [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.897554] env[65385]: WARNING openstack [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.897910] env[65385]: WARNING openstack [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.913017] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afa63664-a416-4b72-97d1-d3658a2c0893 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.920487] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 698.920487] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5246045a-4e24-9d8b-e6df-662ff0bcc37b" [ 698.920487] env[65385]: _type = "Task" [ 698.920487] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.933781] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5246045a-4e24-9d8b-e6df-662ff0bcc37b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.021607] env[65385]: DEBUG nova.network.neutron [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updated VIF entry in instance network info cache for port 59289962-8634-414f-9bd2-b3f5b19af9b4. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 699.023706] env[65385]: DEBUG nova.network.neutron [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.084160] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c764b79-7fe4-4bbc-9f79-5aaf34408d7d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.190s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 699.088041] env[65385]: DEBUG nova.compute.manager [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 699.088255] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.089244] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b762a4e-cc2d-44f4-a985-c4235067437b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.099842] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 699.099842] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27f2115c-a095-4fc1-90bd-a69e7afe0c51 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.107710] env[65385]: DEBUG oslo_vmware.api [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 699.107710] env[65385]: value = "task-4453382" [ 699.107710] env[65385]: _type = "Task" [ 699.107710] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.119769] env[65385]: DEBUG oslo_vmware.api [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.165378] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "34730159-3bec-4ad5-b85e-0f67998d6001" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.165378] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "34730159-3bec-4ad5-b85e-0f67998d6001" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.356659] env[65385]: DEBUG nova.compute.utils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 699.362533] env[65385]: DEBUG nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 699.363405] env[65385]: DEBUG nova.network.neutron [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 699.363405] env[65385]: WARNING neutronclient.v2_0.client [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 699.364421] env[65385]: WARNING neutronclient.v2_0.client [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 699.364511] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.364957] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.436391] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5246045a-4e24-9d8b-e6df-662ff0bcc37b, 'name': SearchDatastore_Task, 'duration_secs': 0.02177} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.440221] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.442193] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.442193] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0e258fe-16f3-4165-877b-c2f67faf272d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.455905] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 699.455905] env[65385]: value = "task-4453383" [ 699.455905] env[65385]: _type = "Task" [ 699.455905] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.468037] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453383, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.489119] env[65385]: DEBUG nova.policy [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de6370f100c842a4875bc6639fbc1f27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f9a3a63deb9427bb5a08c7785509709', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 699.526701] env[65385]: DEBUG oslo_concurrency.lockutils [req-acf1e531-b955-4f1f-9d01-6d808a44bf07 req-1737df3c-e251-40de-821d-b5b7bc711099 service nova] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.564754] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.565100] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.565475] env[65385]: DEBUG nova.compute.manager [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 699.566312] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b2b0fd-239e-43e6-84f3-a26de5998aac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.580503] env[65385]: DEBUG nova.compute.manager [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 699.581604] env[65385]: DEBUG nova.objects.instance [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lazy-loading 'flavor' on Instance uuid 54ae4cd5-76d2-4e76-9528-6f511c4dca23 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 699.623105] env[65385]: DEBUG oslo_vmware.api [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453382, 'name': PowerOffVM_Task, 'duration_secs': 0.287213} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.624321] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 699.624321] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 699.624537] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72cb3d98-ccda-4ddf-a853-ac28b46a12f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.668389] env[65385]: DEBUG nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 699.710544] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 699.710844] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 699.711058] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Deleting the datastore file [datastore1] 8ac44124-d8cf-421c-9ae4-c943df94550a {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 699.711359] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81794407-8d54-4af9-af0d-0535966e01a0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.720710] env[65385]: DEBUG oslo_vmware.api [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for the task: (returnval){ [ 699.720710] env[65385]: value = "task-4453385" [ 699.720710] env[65385]: _type = "Task" [ 699.720710] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.732416] env[65385]: DEBUG oslo_vmware.api [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.861610] env[65385]: DEBUG nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 699.962683] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972de68a-13cb-4265-934e-ede98c5af942 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.982334] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453383, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.985853] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7027cf6-c818-4db5-ae34-2147951df293 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.026466] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5705bc-5bab-46c9-b271-65e85f5923ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.038058] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3428c681-92b0-4c30-9fc1-4b66b3c668fd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.056229] env[65385]: DEBUG nova.compute.provider_tree [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.072419] env[65385]: DEBUG nova.network.neutron [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Successfully created port: 4090f583-8481-4e0c-9a39-3a393fb0ea72 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 700.207130] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.234937] env[65385]: DEBUG oslo_vmware.api [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.236403] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "28fd3828-38ec-426f-b4dc-1e094304780b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.236755] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.475947] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453383, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62488} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.476241] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 700.478024] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.478024] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e29f883-53ad-4ed2-90d0-72bb66bab3cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.485056] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 700.485056] env[65385]: value = "task-4453387" [ 700.485056] env[65385]: _type = "Task" [ 700.485056] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.496243] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453387, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.559923] env[65385]: DEBUG nova.scheduler.client.report [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 700.592028] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 700.592028] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48c1b37b-9177-474f-89a6-b4d292965169 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.601262] env[65385]: DEBUG oslo_vmware.api [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 700.601262] env[65385]: value = "task-4453388" [ 700.601262] env[65385]: _type = "Task" [ 700.601262] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.615860] env[65385]: DEBUG oslo_vmware.api [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453388, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.735670] env[65385]: DEBUG oslo_vmware.api [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Task: {'id': task-4453385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.569657} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.736098] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 700.736250] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 700.736514] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.736773] env[65385]: INFO nova.compute.manager [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Took 1.65 seconds to destroy the instance on the hypervisor. [ 700.737158] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 700.737427] env[65385]: DEBUG nova.compute.manager [-] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 700.737554] env[65385]: DEBUG nova.network.neutron [-] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 700.737852] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.739158] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.739536] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.749053] env[65385]: DEBUG nova.compute.utils [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 700.872394] env[65385]: DEBUG nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 700.894720] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 700.895055] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 700.895228] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 700.895409] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 700.895569] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 700.895711] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 700.895921] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.896177] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 700.896354] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 700.896509] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 700.896693] env[65385]: DEBUG nova.virt.hardware [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 700.897931] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f46fba-e2d7-490c-8ff8-4d0c931173dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.907250] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97dfb8e3-61de-4c2f-bb99-d06c53f0693f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.966246] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.995899] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453387, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074354} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.996258] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 700.997105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b116b09a-cadf-4227-9152-56d7aa93e59c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.020605] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.020949] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58eb4c3f-dd92-4fab-9fff-c82ee8d1c8b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.044084] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 701.044084] env[65385]: value = "task-4453389" [ 701.044084] env[65385]: _type = "Task" [ 701.044084] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.054648] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453389, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.065189] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.212s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.068028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 26.826s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.109075] env[65385]: INFO nova.scheduler.client.report [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Deleted allocations for instance bf79fbcd-e943-4ff3-bcf8-6a2817df8292 [ 701.116599] env[65385]: DEBUG oslo_vmware.api [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453388, 'name': PowerOffVM_Task, 'duration_secs': 0.465523} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.117289] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 701.117366] env[65385]: DEBUG nova.compute.manager [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 701.118162] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb14318-d14d-4b80-904b-4f5fd67010eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.251829] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.015s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.555973] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.621600] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b7b24f61-c866-4ff0-8816-19a55cb51e55 tempest-FloatingIPsAssociationNegativeTestJSON-118792969 tempest-FloatingIPsAssociationNegativeTestJSON-118792969-project-member] Lock "bf79fbcd-e943-4ff3-bcf8-6a2817df8292" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.236s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.633786] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fbb58da-587a-4fff-9734-75afee52e6e1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.069s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.706698] env[65385]: DEBUG nova.network.neutron [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Successfully updated port: 4090f583-8481-4e0c-9a39-3a393fb0ea72 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 702.057028] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453389, 'name': ReconfigVM_Task, 'duration_secs': 0.863973} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.057028] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Reconfigured VM instance instance-0000001d to attach disk [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9/8b478626-e3dc-42c4-bd59-fa0a71d798c9.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.057547] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdcaccdc-b882-4f90-be91-a5895d6ad847 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.066234] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 702.066234] env[65385]: value = "task-4453390" [ 702.066234] env[65385]: _type = "Task" [ 702.066234] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.075872] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453390, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.108387] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 8ac44124-d8cf-421c-9ae4-c943df94550a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.108387] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 649b4e21-fc55-415c-a1f3-ec724397b874 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.108387] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0115dde8-b0aa-4960-a452-ced6587bc567 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.108387] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.108387] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 60067529-1071-4295-b1c9-21523bf347d9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 702.108387] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 59450963-d83f-46e8-8c13-05d4f1818c64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.108387] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 1bc94ee7-d4f9-48c3-97f1-9e662eb50582 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.108387] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 427d5c6d-ab01-42ef-8ee5-edd608896b8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.108732] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance eb9d0510-b453-4695-9e1d-731217b9f8ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.109830] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 42b9edf6-1873-49c3-8074-8eef654ac371 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.109830] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 75ba706f-2e72-4f84-b02f-db4381951e77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.109830] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 69fdd5df-a9f2-486f-8a79-87c034366083 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 702.109830] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance de8632e0-bf38-440d-b6a3-895efeef122d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 702.109830] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 5291c04c-24ad-4c64-9fc2-fde42da4bc0f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 702.109830] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 89f3c815-8671-47ce-9e74-bf6e652bb3c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.110266] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance a71d2fa6-006c-4f76-8ae8-467e149297da is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 702.110514] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 950efcc6-7e4f-4b30-b0fa-d940f893e1d5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 702.110698] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 2e4c064f-27b0-4ccb-8fe6-c34a61153663 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.111875] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 8b478626-e3dc-42c4-bd59-fa0a71d798c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.112035] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 59f26752-75ab-45e5-b0a9-bbcef44ecd7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.112140] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0673a0d2-76ac-4cd5-8cc3-b8596877c641 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.112294] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 54ae4cd5-76d2-4e76-9528-6f511c4dca23 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.112349] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 28fd3828-38ec-426f-b4dc-1e094304780b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.112449] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 3657ec21-3ec7-44c9-92d1-570655b58f26 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 702.135841] env[65385]: DEBUG nova.network.neutron [-] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 702.210259] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "refresh_cache-3657ec21-3ec7-44c9-92d1-570655b58f26" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.210499] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired lock "refresh_cache-3657ec21-3ec7-44c9-92d1-570655b58f26" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.210748] env[65385]: DEBUG nova.network.neutron [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 702.307120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "28fd3828-38ec-426f-b4dc-1e094304780b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.307120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.308805] env[65385]: INFO nova.compute.manager [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Attaching volume 4d7ca81e-377c-4dd0-bf1f-39f6447b0eda to /dev/sdb [ 702.382299] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880597f0-cfb6-4f07-8b28-066ced456165 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.396690] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f9a124-539d-42e9-958b-0842ad478265 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.415654] env[65385]: DEBUG nova.virt.block_device [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Updating existing volume attachment record: 65c2bfb4-420c-4eab-ad79-8cd220b4d6cf {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 702.468181] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.468415] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.542488] env[65385]: DEBUG nova.compute.manager [req-3a09d585-d8fc-4ce9-afff-957ca076b62c req-52f0b7a8-4ab1-4650-8c65-b27e8b69caa3 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Received event network-vif-plugged-4090f583-8481-4e0c-9a39-3a393fb0ea72 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 702.543143] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a09d585-d8fc-4ce9-afff-957ca076b62c req-52f0b7a8-4ab1-4650-8c65-b27e8b69caa3 service nova] Acquiring lock "3657ec21-3ec7-44c9-92d1-570655b58f26-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.543143] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a09d585-d8fc-4ce9-afff-957ca076b62c req-52f0b7a8-4ab1-4650-8c65-b27e8b69caa3 service nova] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.543275] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a09d585-d8fc-4ce9-afff-957ca076b62c req-52f0b7a8-4ab1-4650-8c65-b27e8b69caa3 service nova] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.543483] env[65385]: DEBUG nova.compute.manager [req-3a09d585-d8fc-4ce9-afff-957ca076b62c req-52f0b7a8-4ab1-4650-8c65-b27e8b69caa3 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] No waiting events found dispatching network-vif-plugged-4090f583-8481-4e0c-9a39-3a393fb0ea72 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 702.543762] env[65385]: WARNING nova.compute.manager [req-3a09d585-d8fc-4ce9-afff-957ca076b62c req-52f0b7a8-4ab1-4650-8c65-b27e8b69caa3 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Received unexpected event network-vif-plugged-4090f583-8481-4e0c-9a39-3a393fb0ea72 for instance with vm_state building and task_state spawning. [ 702.578727] env[65385]: DEBUG nova.compute.manager [req-61d80305-07c5-43c8-86cd-72abc067bd1d req-019a9ea3-512d-4e9c-adc3-a39746273715 service nova] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Received event network-vif-deleted-f5dabab4-c5af-42f4-bfdc-eec4330bdb0d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 702.586880] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453390, 'name': Rename_Task, 'duration_secs': 0.277981} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.587251] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.587586] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb4ec300-004c-4d71-8c70-00141e24255a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.596350] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 702.596350] env[65385]: value = "task-4453393" [ 702.596350] env[65385]: _type = "Task" [ 702.596350] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.608871] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453393, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.617918] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 90c36f67-0c25-4487-a5e4-89827290953b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 702.637760] env[65385]: INFO nova.compute.manager [-] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Took 1.90 seconds to deallocate network for instance. [ 702.716647] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.717110] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.776811] env[65385]: DEBUG nova.network.neutron [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 702.964934] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.965455] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.973972] env[65385]: DEBUG nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 703.111139] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453393, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.121372] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 2f66a03b-874b-4bee-9694-49813ecb0c8a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 703.145575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.261475] env[65385]: WARNING neutronclient.v2_0.client [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.262258] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.262771] env[65385]: WARNING openstack [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.493863] env[65385]: DEBUG nova.network.neutron [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Updating instance_info_cache with network_info: [{"id": "4090f583-8481-4e0c-9a39-3a393fb0ea72", "address": "fa:16:3e:17:a1:1b", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4090f583-84", "ovs_interfaceid": "4090f583-8481-4e0c-9a39-3a393fb0ea72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.511479] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.608889] env[65385]: DEBUG oslo_vmware.api [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453393, 'name': PowerOnVM_Task, 'duration_secs': 0.5893} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.609201] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 703.609411] env[65385]: DEBUG nova.compute.manager [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 703.610291] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18cf866-178f-4177-9c81-5774ad8c9589 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.626457] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 483cf486-f0fc-48a4-9db2-970d0f35865a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 704.000576] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Releasing lock "refresh_cache-3657ec21-3ec7-44c9-92d1-570655b58f26" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.000576] env[65385]: DEBUG nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Instance network_info: |[{"id": "4090f583-8481-4e0c-9a39-3a393fb0ea72", "address": "fa:16:3e:17:a1:1b", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4090f583-84", "ovs_interfaceid": "4090f583-8481-4e0c-9a39-3a393fb0ea72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 704.000576] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:a1:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4090f583-8481-4e0c-9a39-3a393fb0ea72', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 704.010432] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Creating folder: Project (5f9a3a63deb9427bb5a08c7785509709). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 704.011589] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3fcb8666-8f48-4bf0-b2ae-7bc5d46d8145 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.025078] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Created folder: Project (5f9a3a63deb9427bb5a08c7785509709) in parent group-v870881. [ 704.025078] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Creating folder: Instances. Parent ref: group-v870983. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 704.025380] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0118ad17-9ab4-4544-85be-143db88ebb81 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.038106] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Created folder: Instances in parent group-v870983. [ 704.038391] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 704.038601] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 704.038847] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bacc3913-5bc6-465b-8559-e82464640c63 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.060889] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.060889] env[65385]: value = "task-4453398" [ 704.060889] env[65385]: _type = "Task" [ 704.060889] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.070138] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453398, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.129732] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 7fcbdc79-688c-479f-94e2-f4542abe714b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 704.133116] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.221013] env[65385]: DEBUG nova.compute.manager [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 704.222034] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbff0d5-fbfc-4649-9b39-8d141b1fa1da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.576743] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453398, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.614404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.614686] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.614892] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.615166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.615437] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.620120] env[65385]: INFO nova.compute.manager [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Terminating instance [ 704.635791] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance c798006c-3e74-461f-8397-3d8ca0be0a95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 704.735391] env[65385]: INFO nova.compute.manager [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] instance snapshotting [ 704.735391] env[65385]: WARNING nova.compute.manager [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 704.737757] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae3ff46-2a1d-4bfe-9363-1113bf86f6c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.758877] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51428043-e713-4a51-9f0f-147519176483 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.778274] env[65385]: DEBUG nova.compute.manager [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Received event network-changed-4090f583-8481-4e0c-9a39-3a393fb0ea72 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 704.778527] env[65385]: DEBUG nova.compute.manager [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Refreshing instance network info cache due to event network-changed-4090f583-8481-4e0c-9a39-3a393fb0ea72. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 704.779122] env[65385]: DEBUG oslo_concurrency.lockutils [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Acquiring lock "refresh_cache-3657ec21-3ec7-44c9-92d1-570655b58f26" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.779122] env[65385]: DEBUG oslo_concurrency.lockutils [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Acquired lock "refresh_cache-3657ec21-3ec7-44c9-92d1-570655b58f26" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.779218] env[65385]: DEBUG nova.network.neutron [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Refreshing network info cache for port 4090f583-8481-4e0c-9a39-3a393fb0ea72 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 705.073224] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453398, 'name': CreateVM_Task, 'duration_secs': 0.522642} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.073477] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 705.074383] env[65385]: WARNING neutronclient.v2_0.client [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.075447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.075447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.075447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 705.075604] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-376e541b-c867-4dd6-9bf6-e8eefd4c2f51 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.082034] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 705.082034] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522af630-88a8-27b1-39af-2641b0c8cb7d" [ 705.082034] env[65385]: _type = "Task" [ 705.082034] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.092264] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522af630-88a8-27b1-39af-2641b0c8cb7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.126736] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "refresh_cache-8b478626-e3dc-42c4-bd59-fa0a71d798c9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.126736] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquired lock "refresh_cache-8b478626-e3dc-42c4-bd59-fa0a71d798c9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.126736] env[65385]: DEBUG nova.network.neutron [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 705.139036] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 34730159-3bec-4ad5-b85e-0f67998d6001 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 705.271528] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 705.272019] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dc36db7a-c7e8-423c-a6c8-bbbda894cc45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.284121] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 705.284121] env[65385]: value = "task-4453401" [ 705.284121] env[65385]: _type = "Task" [ 705.284121] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.285060] env[65385]: WARNING neutronclient.v2_0.client [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.286118] env[65385]: WARNING openstack [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.286687] env[65385]: WARNING openstack [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.308166] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453401, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.453584] env[65385]: WARNING openstack [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.453951] env[65385]: WARNING openstack [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.534489] env[65385]: WARNING neutronclient.v2_0.client [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.535248] env[65385]: WARNING openstack [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.535614] env[65385]: WARNING openstack [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.601409] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522af630-88a8-27b1-39af-2641b0c8cb7d, 'name': SearchDatastore_Task, 'duration_secs': 0.020527} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.609398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.609398] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 705.609398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.609398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.609398] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 705.609746] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecc590ea-f7ed-488e-96e8-55ced8bfddc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.625254] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 705.625254] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 705.625254] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab5f3401-9383-40b3-99a2-e4af8d5c0254 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.630352] env[65385]: WARNING neutronclient.v2_0.client [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.631720] env[65385]: WARNING openstack [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.632263] env[65385]: WARNING openstack [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.641707] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 705.642353] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 705.643001] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4096MB phys_disk=100GB used_disk=18GB total_vcpus=48 used_vcpus=18 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '18', 'num_vm_active': '17', 'num_task_None': '14', 'num_os_type_None': '18', 'num_proj_99173b1f52aa42f9b0db92366db8c411': '1', 'io_workload': '1', 'num_task_deleting': '1', 'num_proj_349d3e2d0d8849819007c19528145c3c': '1', 'num_proj_112402342b0e481f83be7e17c9f739e9': '3', 'num_proj_e94f2a868e2d4d31af54f2e5d8c479f8': '1', 'num_proj_f858126fa23d43cbafd1b677206f28ac': '1', 'num_proj_2de8098800694299aae5aa2f59f405bb': '1', 'num_proj_a9a1bf7c2dad4808a3782a3fba8a25fc': '3', 'num_proj_1598307c304d46a884a88b3eb5f5d7dc': '1', 'num_proj_c8167324e3664a578a5ca8c8a04c1241': '1', 'num_task_rebuild_spawning': '1', 'num_proj_0e778e71f7644764b8d0d87d973280eb': '1', 'num_proj_60de0d1c162342209795a180391661f6': '1', 'num_task_powering-off': '1', 'num_proj_c27eb6232a2148a1a259f57494b4ae30': '1', 'num_proj_754b1facaaa14501b2204c98e1d7a5a8': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_5f9a3a63deb9427bb5a08c7785509709': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 705.650377] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 705.650377] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c54f5f-ada0-3d0b-683d-e178fc533dcc" [ 705.650377] env[65385]: _type = "Task" [ 705.650377] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.665590] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c54f5f-ada0-3d0b-683d-e178fc533dcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.669453] env[65385]: DEBUG nova.network.neutron [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Updated VIF entry in instance network info cache for port 4090f583-8481-4e0c-9a39-3a393fb0ea72. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 705.670232] env[65385]: DEBUG nova.network.neutron [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Updating instance_info_cache with network_info: [{"id": "4090f583-8481-4e0c-9a39-3a393fb0ea72", "address": "fa:16:3e:17:a1:1b", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4090f583-84", "ovs_interfaceid": "4090f583-8481-4e0c-9a39-3a393fb0ea72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 705.673222] env[65385]: DEBUG nova.network.neutron [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 705.755446] env[65385]: DEBUG nova.network.neutron [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 705.800802] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453401, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.174533] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.174758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.175935] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c54f5f-ada0-3d0b-683d-e178fc533dcc, 'name': SearchDatastore_Task, 'duration_secs': 0.020396} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.176381] env[65385]: DEBUG oslo_concurrency.lockutils [req-bad7d1d2-27ee-46a8-bf4c-f847945b9ec5 req-174da3ff-58b4-4955-a536-b5acad5bfb69 service nova] Releasing lock "refresh_cache-3657ec21-3ec7-44c9-92d1-570655b58f26" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 706.177592] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9ab92eb-5348-4efe-bbf8-05d8a09058ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.190239] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 706.190239] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d850cd-4ab9-cba0-29f3-0f5916949df4" [ 706.190239] env[65385]: _type = "Task" [ 706.190239] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.209450] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d850cd-4ab9-cba0-29f3-0f5916949df4, 'name': SearchDatastore_Task, 'duration_secs': 0.015487} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.209623] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 706.209746] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 3657ec21-3ec7-44c9-92d1-570655b58f26/3657ec21-3ec7-44c9-92d1-570655b58f26.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 706.210043] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42a70006-844f-4d52-bd07-f66eb65939d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.215860] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0babcc87-a2e3-4581-b5dd-36e7800cb44a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.226890] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482d6ce3-138d-4903-8f00-911748af2d7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.230395] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 706.230395] env[65385]: value = "task-4453402" [ 706.230395] env[65385]: _type = "Task" [ 706.230395] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.261754] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Releasing lock "refresh_cache-8b478626-e3dc-42c4-bd59-fa0a71d798c9" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 706.261754] env[65385]: DEBUG nova.compute.manager [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 706.262046] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.264159] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4a5680-34ee-43e3-b793-1dee4e1834ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.272896] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d27d10c-ed29-4f81-bd2d-4e27da616edb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.284042] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453402, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.291931] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 706.295690] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c40f5f32-6496-497f-b7ba-48f1f962fd0c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.301025] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50f3f56-c057-4370-b58c-d630b6ce13e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.309582] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453401, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.323532] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.329025] env[65385]: DEBUG oslo_vmware.api [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 706.329025] env[65385]: value = "task-4453403" [ 706.329025] env[65385]: _type = "Task" [ 706.329025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.339333] env[65385]: DEBUG oslo_vmware.api [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.453959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.453959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.453959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.453959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.453959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.454806] env[65385]: INFO nova.compute.manager [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Terminating instance [ 706.684886] env[65385]: DEBUG nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 706.745942] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453402, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.803401] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453401, 'name': CreateSnapshot_Task, 'duration_secs': 1.227788} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.803702] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 706.804471] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4932d039-a521-4cc3-a97b-1025f6ae1c7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.829106] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 706.846353] env[65385]: DEBUG oslo_vmware.api [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453403, 'name': PowerOffVM_Task, 'duration_secs': 0.218737} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.846353] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 706.846353] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 706.846353] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-acefa717-478e-4964-b44f-e6f88454555f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.879022] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 706.879022] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 706.879022] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Deleting the datastore file [datastore1] 8b478626-e3dc-42c4-bd59-fa0a71d798c9 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.879387] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-591081eb-b2ba-486a-ab1d-84457dd9dbfa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.887865] env[65385]: DEBUG oslo_vmware.api [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for the task: (returnval){ [ 706.887865] env[65385]: value = "task-4453406" [ 706.887865] env[65385]: _type = "Task" [ 706.887865] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.900022] env[65385]: DEBUG oslo_vmware.api [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453406, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.962740] env[65385]: DEBUG nova.compute.manager [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 706.963048] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.964080] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d3a461-eedf-4c02-9eee-2e35b37cf817 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.972872] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 706.973171] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ab82259-f368-4298-9d07-71136e58f7a1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.981617] env[65385]: DEBUG oslo_vmware.api [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 706.981617] env[65385]: value = "task-4453407" [ 706.981617] env[65385]: _type = "Task" [ 706.981617] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.992349] env[65385]: DEBUG oslo_vmware.api [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453407, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.222545] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.242224] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453402, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685696} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.242505] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 3657ec21-3ec7-44c9-92d1-570655b58f26/3657ec21-3ec7-44c9-92d1-570655b58f26.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 707.242871] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 707.242959] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d90e9328-d119-4d77-adc4-aa3f0a593c38 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.251804] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 707.251804] env[65385]: value = "task-4453408" [ 707.251804] env[65385]: _type = "Task" [ 707.251804] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.262372] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453408, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.328843] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 707.329271] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-31769179-48b8-4f63-8c97-569cb6ae2aa8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.337432] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 707.337758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.270s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.337871] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.142s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 707.338063] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.340290] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.494s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 707.340500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.342401] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.227s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 707.344094] env[65385]: INFO nova.compute.claims [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.350055] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 707.350055] env[65385]: value = "task-4453409" [ 707.350055] env[65385]: _type = "Task" [ 707.350055] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.359871] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453409, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.376047] env[65385]: INFO nova.scheduler.client.report [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Deleted allocations for instance de8632e0-bf38-440d-b6a3-895efeef122d [ 707.380344] env[65385]: INFO nova.scheduler.client.report [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Deleted allocations for instance 69fdd5df-a9f2-486f-8a79-87c034366083 [ 707.401418] env[65385]: DEBUG oslo_vmware.api [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Task: {'id': task-4453406, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162674} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.401700] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.401877] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 707.402054] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.402252] env[65385]: INFO nova.compute.manager [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 707.402536] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 707.402755] env[65385]: DEBUG nova.compute.manager [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 707.402969] env[65385]: DEBUG nova.network.neutron [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 707.403301] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.403850] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.404127] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.439793] env[65385]: DEBUG nova.network.neutron [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 707.440106] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.488947] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 707.489249] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870982', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'name': 'volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '28fd3828-38ec-426f-b4dc-1e094304780b', 'attached_at': '', 'detached_at': '', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'serial': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 707.490240] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83f0c38-aab3-4b50-9246-81b26bfce44b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.499791] env[65385]: DEBUG oslo_vmware.api [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453407, 'name': PowerOffVM_Task, 'duration_secs': 0.225373} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.511529] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 707.511874] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 707.512403] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3fb68cb5-c212-4e6a-91ee-c28ee24ae150 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.514828] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd90796-c612-450e-ac26-601fdb7800ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.545306] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda/volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.545306] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-055652de-3bcf-4b3e-8778-b22f608a4471 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.564893] env[65385]: DEBUG oslo_vmware.api [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 707.564893] env[65385]: value = "task-4453411" [ 707.564893] env[65385]: _type = "Task" [ 707.564893] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.574870] env[65385]: DEBUG oslo_vmware.api [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453411, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.608736] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 707.609257] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 707.609467] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleting the datastore file [datastore2] 1bc94ee7-d4f9-48c3-97f1-9e662eb50582 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 707.610148] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a50cc97-68fe-4f8c-9154-8b49baa64391 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.617701] env[65385]: DEBUG oslo_vmware.api [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 707.617701] env[65385]: value = "task-4453412" [ 707.617701] env[65385]: _type = "Task" [ 707.617701] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.627863] env[65385]: DEBUG oslo_vmware.api [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.763767] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453408, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089543} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.764133] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 707.765188] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbef9db-a869-42ec-be57-864cb069e061 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.792626] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 3657ec21-3ec7-44c9-92d1-570655b58f26/3657ec21-3ec7-44c9-92d1-570655b58f26.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.793313] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d964af55-3cee-4388-9ce7-3f285b9e2301 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.815930] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 707.815930] env[65385]: value = "task-4453413" [ 707.815930] env[65385]: _type = "Task" [ 707.815930] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.827133] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453413, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.863796] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453409, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.886937] env[65385]: DEBUG oslo_concurrency.lockutils [None req-49d7f403-6b3f-4592-9edd-46c97c0ac17d tempest-InstanceActionsNegativeTestJSON-206719615 tempest-InstanceActionsNegativeTestJSON-206719615-project-member] Lock "de8632e0-bf38-440d-b6a3-895efeef122d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.543s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.888152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c4f5e19-39a0-4a67-94e3-9a6b104b1740 tempest-DeleteServersAdminTestJSON-1286097210 tempest-DeleteServersAdminTestJSON-1286097210-project-member] Lock "69fdd5df-a9f2-486f-8a79-87c034366083" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.194s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.942564] env[65385]: DEBUG nova.network.neutron [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 708.076248] env[65385]: DEBUG oslo_vmware.api [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453411, 'name': ReconfigVM_Task, 'duration_secs': 0.491558} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.076505] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Reconfigured VM instance instance-00000021 to attach disk [datastore1] volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda/volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.082534] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c99a5bbb-9112-49c5-9b41-9987586525d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.100056] env[65385]: DEBUG oslo_vmware.api [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 708.100056] env[65385]: value = "task-4453414" [ 708.100056] env[65385]: _type = "Task" [ 708.100056] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.110605] env[65385]: DEBUG oslo_vmware.api [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453414, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.129542] env[65385]: DEBUG oslo_vmware.api [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191075} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.129791] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.129988] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 708.130212] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.130469] env[65385]: INFO nova.compute.manager [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Took 1.17 seconds to destroy the instance on the hypervisor. [ 708.130741] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 708.131041] env[65385]: DEBUG nova.compute.manager [-] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 708.131041] env[65385]: DEBUG nova.network.neutron [-] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 708.131440] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.131994] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.132300] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.218851] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.329949] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453413, 'name': ReconfigVM_Task, 'duration_secs': 0.30517} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.330526] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 3657ec21-3ec7-44c9-92d1-570655b58f26/3657ec21-3ec7-44c9-92d1-570655b58f26.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.331291] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b149353-48af-4439-8e45-2f692a9561ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.342273] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 708.342273] env[65385]: value = "task-4453415" [ 708.342273] env[65385]: _type = "Task" [ 708.342273] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.356219] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453415, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.371820] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453409, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.447496] env[65385]: INFO nova.compute.manager [-] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Took 1.04 seconds to deallocate network for instance. [ 708.613992] env[65385]: DEBUG oslo_vmware.api [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453414, 'name': ReconfigVM_Task, 'duration_secs': 0.152117} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.614620] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870982', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'name': 'volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '28fd3828-38ec-426f-b4dc-1e094304780b', 'attached_at': '', 'detached_at': '', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'serial': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 708.747177] env[65385]: DEBUG nova.compute.manager [req-bc800a7a-3c4d-4094-8e38-fe4df0175b9b req-f08d340a-1fb1-4dde-8c8d-891b9b6cfed9 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Received event network-vif-deleted-5f60e5e9-2860-4364-99c0-3c1433e4bddb {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 708.747440] env[65385]: INFO nova.compute.manager [req-bc800a7a-3c4d-4094-8e38-fe4df0175b9b req-f08d340a-1fb1-4dde-8c8d-891b9b6cfed9 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Neutron deleted interface 5f60e5e9-2860-4364-99c0-3c1433e4bddb; detaching it from the instance and deleting it from the info cache [ 708.747739] env[65385]: DEBUG nova.network.neutron [req-bc800a7a-3c4d-4094-8e38-fe4df0175b9b req-f08d340a-1fb1-4dde-8c8d-891b9b6cfed9 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 708.853597] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453415, 'name': Rename_Task, 'duration_secs': 0.163373} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.856590] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 708.857833] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8843badb-d6d4-438f-96fd-17d52626129b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.868335] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453409, 'name': CloneVM_Task, 'duration_secs': 1.253933} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.872443] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Created linked-clone VM from snapshot [ 708.872849] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 708.872849] env[65385]: value = "task-4453417" [ 708.872849] env[65385]: _type = "Task" [ 708.872849] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.873826] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc5f302-030d-41ae-b124-d2ea1e80d46c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.886940] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453417, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.893651] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Uploading image 73727e37-3c18-44bd-8f2d-ba607b985a7a {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 708.936417] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 708.936417] env[65385]: value = "vm-870987" [ 708.936417] env[65385]: _type = "VirtualMachine" [ 708.936417] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 708.936825] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c1790843-fab6-4f16-921d-8e87e34c1b1a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.948990] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease: (returnval){ [ 708.948990] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52240457-1603-0e41-0158-589feef3892b" [ 708.948990] env[65385]: _type = "HttpNfcLease" [ 708.948990] env[65385]: } obtained for exporting VM: (result){ [ 708.948990] env[65385]: value = "vm-870987" [ 708.948990] env[65385]: _type = "VirtualMachine" [ 708.948990] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 708.948990] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the lease: (returnval){ [ 708.948990] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52240457-1603-0e41-0158-589feef3892b" [ 708.948990] env[65385]: _type = "HttpNfcLease" [ 708.948990] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 708.957096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.957393] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 708.957393] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52240457-1603-0e41-0158-589feef3892b" [ 708.957393] env[65385]: _type = "HttpNfcLease" [ 708.957393] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 709.022751] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec26be4e-9c61-4999-ae99-31e1e466bf20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.031040] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5dcfdd-cffa-4b15-a0a3-51bad597e295 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.034786] env[65385]: DEBUG nova.network.neutron [-] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 709.072862] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f2345d-4647-4689-8161-24e9b8718aa7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.084191] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92575a2e-f8c2-45ad-bd12-4b078c3aadd8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.100199] env[65385]: DEBUG nova.compute.provider_tree [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.256994] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eed4602f-3f31-4eda-ac9a-e3659c7bcd06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.273919] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a9571f-3d60-4e7b-8aaf-530a67240de7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.324182] env[65385]: DEBUG nova.compute.manager [req-bc800a7a-3c4d-4094-8e38-fe4df0175b9b req-f08d340a-1fb1-4dde-8c8d-891b9b6cfed9 service nova] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Detach interface failed, port_id=5f60e5e9-2860-4364-99c0-3c1433e4bddb, reason: Instance 1bc94ee7-d4f9-48c3-97f1-9e662eb50582 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 709.387723] env[65385]: DEBUG oslo_vmware.api [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453417, 'name': PowerOnVM_Task, 'duration_secs': 0.492441} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.387944] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 709.388194] env[65385]: INFO nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Took 8.52 seconds to spawn the instance on the hypervisor. [ 709.388328] env[65385]: DEBUG nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 709.389143] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cece4dca-7797-41b5-a07d-9493638cc9bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.455929] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 709.455929] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52240457-1603-0e41-0158-589feef3892b" [ 709.455929] env[65385]: _type = "HttpNfcLease" [ 709.455929] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 709.456687] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 709.456687] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52240457-1603-0e41-0158-589feef3892b" [ 709.456687] env[65385]: _type = "HttpNfcLease" [ 709.456687] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 709.457712] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e600eea5-e8bc-4239-a89c-6a1425878dc8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.467429] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52995cc4-34aa-1d5c-69ab-80111d1cbf79/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 709.467661] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52995cc4-34aa-1d5c-69ab-80111d1cbf79/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 709.538835] env[65385]: INFO nova.compute.manager [-] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Took 1.41 seconds to deallocate network for instance. [ 709.604070] env[65385]: DEBUG nova.scheduler.client.report [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 709.642454] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7fb1b533-e504-45be-ad66-6de0c240452d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.683306] env[65385]: DEBUG nova.objects.instance [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'flavor' on Instance uuid 28fd3828-38ec-426f-b4dc-1e094304780b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 709.798221] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.798467] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.798767] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.798873] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.798993] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.802128] env[65385]: INFO nova.compute.manager [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Terminating instance [ 709.912856] env[65385]: INFO nova.compute.manager [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Took 36.36 seconds to build instance. [ 710.048343] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.112382] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.770s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.113130] env[65385]: DEBUG nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 710.119851] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.999s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.122693] env[65385]: INFO nova.compute.claims [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.196760] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fba34010-0c2e-4926-bf32-44d920c113f2 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.890s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.307817] env[65385]: DEBUG nova.compute.manager [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 710.308284] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 710.309220] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258e7d97-ff5c-4c36-a34c-ab0b4423e51e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.320965] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 710.321489] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf6ddd10-1451-48a2-a42f-12d3ee322a67 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.331790] env[65385]: DEBUG oslo_vmware.api [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 710.331790] env[65385]: value = "task-4453419" [ 710.331790] env[65385]: _type = "Task" [ 710.331790] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.349081] env[65385]: DEBUG oslo_vmware.api [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453419, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.413347] env[65385]: DEBUG oslo_concurrency.lockutils [None req-186270c1-4f95-4aaa-9d9f-3edcd819b995 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.907s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.641024] env[65385]: DEBUG nova.compute.utils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 710.641802] env[65385]: DEBUG nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 710.642054] env[65385]: DEBUG nova.network.neutron [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 710.642387] env[65385]: WARNING neutronclient.v2_0.client [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 710.645592] env[65385]: WARNING neutronclient.v2_0.client [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 710.645592] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.645592] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.714683] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "fae095a8-a768-4263-b456-ed0745398755" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.715834] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "fae095a8-a768-4263-b456-ed0745398755" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.776564] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.776741] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.794502] env[65385]: DEBUG nova.policy [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18c6274d42a04fbe81729aad5faf8a98', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cb1ecd79d8e44f6096ca69a5238851b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 710.848485] env[65385]: DEBUG oslo_vmware.api [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453419, 'name': PowerOffVM_Task, 'duration_secs': 0.237244} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.848885] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 710.849056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 710.849396] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4facbd20-13aa-4860-9521-1aa1d106143e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.921532] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 710.921867] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 710.921933] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Deleting the datastore file [datastore2] 427d5c6d-ab01-42ef-8ee5-edd608896b8b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 710.922373] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c46005b2-0712-40dd-bec6-16c3a34c3b4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.930690] env[65385]: DEBUG oslo_vmware.api [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 710.930690] env[65385]: value = "task-4453421" [ 710.930690] env[65385]: _type = "Task" [ 710.930690] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.940197] env[65385]: DEBUG oslo_vmware.api [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453421, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.063619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "28fd3828-38ec-426f-b4dc-1e094304780b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.063619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.063619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "28fd3828-38ec-426f-b4dc-1e094304780b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.063619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.063895] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.066387] env[65385]: INFO nova.compute.manager [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Terminating instance [ 711.159023] env[65385]: DEBUG nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 711.220329] env[65385]: DEBUG nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 711.255121] env[65385]: DEBUG nova.network.neutron [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Successfully created port: c2cde5c4-0094-40e7-965c-60977c968a4d {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 711.280687] env[65385]: DEBUG nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 711.457973] env[65385]: DEBUG oslo_vmware.api [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453421, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212772} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.457973] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 711.457973] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 711.457973] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.460018] env[65385]: INFO nova.compute.manager [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 711.460018] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 711.460018] env[65385]: DEBUG nova.compute.manager [-] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 711.460018] env[65385]: DEBUG nova.network.neutron [-] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 711.460018] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.460018] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.460804] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.572922] env[65385]: DEBUG nova.compute.manager [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 711.573349] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 711.573733] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24427cd6-88c2-4a5c-bd80-375a55ad4dd5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.582705] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 711.582705] env[65385]: value = "task-4453422" [ 711.582705] env[65385]: _type = "Task" [ 711.582705] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.596320] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.750979] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.808302] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.825512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f7c7e5-6889-409f-88c3-b132752bf573 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.834446] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c2c0ee-dff0-478d-8db5-ad5e48505de1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.866264] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68169f9-6881-4484-8e4d-89a4759ad73c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.875729] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113123d8-cc75-4e4a-8234-91d4136b1083 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.893694] env[65385]: DEBUG nova.compute.provider_tree [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.082507] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.102088] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453422, 'name': PowerOffVM_Task, 'duration_secs': 0.225247} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.102590] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 712.102901] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 712.103336] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870982', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'name': 'volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '28fd3828-38ec-426f-b4dc-1e094304780b', 'attached_at': '', 'detached_at': '', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'serial': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 712.104957] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2a8a22-db99-4244-9606-2eda2ef5792f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.132600] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d204fea4-ab51-4286-a26b-b48ddcd7b31d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.142158] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230bf797-fd7a-4852-b1fd-9c9a7d68ad99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.168798] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795ba28e-6c0e-4a2c-8f1e-52c650a84ff2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.172642] env[65385]: DEBUG nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 712.191363] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The volume has not been displaced from its original location: [datastore1] volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda/volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 712.196779] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Reconfiguring VM instance instance-00000021 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 712.199327] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3b0f417-1024-47be-9cf2-e729bd0c94b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.220103] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 712.220103] env[65385]: value = "task-4453423" [ 712.220103] env[65385]: _type = "Task" [ 712.220103] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.222707] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 712.222927] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 712.223103] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 712.223321] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 712.223474] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 712.223607] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 712.223815] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.223970] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 712.224146] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 712.224304] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 712.224483] env[65385]: DEBUG nova.virt.hardware [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 712.225505] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2258a3f5-9ea8-44dc-b2d9-0a72edefad51 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.240776] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3549562a-de0b-41b8-85bb-e1826b80d811 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.245388] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453423, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.397255] env[65385]: DEBUG nova.scheduler.client.report [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 712.736615] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453423, 'name': ReconfigVM_Task, 'duration_secs': 0.296848} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.737412] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Reconfigured VM instance instance-00000021 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 712.743754] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88db6d7c-82f7-408e-b43e-a80ec08c3833 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.769095] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 712.769095] env[65385]: value = "task-4453424" [ 712.769095] env[65385]: _type = "Task" [ 712.769095] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.781222] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453424, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.905174] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.788s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.905174] env[65385]: DEBUG nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 712.911613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.291s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.912578] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.004s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.918727] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.669s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.920131] env[65385]: INFO nova.compute.claims [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.959325] env[65385]: INFO nova.scheduler.client.report [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Deleted allocations for instance a71d2fa6-006c-4f76-8ae8-467e149297da [ 712.984898] env[65385]: DEBUG nova.network.neutron [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Successfully updated port: c2cde5c4-0094-40e7-965c-60977c968a4d {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 713.280345] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453424, 'name': ReconfigVM_Task, 'duration_secs': 0.353508} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.280692] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870982', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'name': 'volume-4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '28fd3828-38ec-426f-b4dc-1e094304780b', 'attached_at': '', 'detached_at': '', 'volume_id': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda', 'serial': '4d7ca81e-377c-4dd0-bf1f-39f6447b0eda'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 713.281100] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 713.281816] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bed374-4d27-4d60-a9f0-c06f985f56a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.294386] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 713.294386] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-241e8d39-5009-4730-9e1f-c5204cd54ff0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.376758] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 713.377443] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 713.377564] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleting the datastore file [datastore1] 28fd3828-38ec-426f-b4dc-1e094304780b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 713.377806] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7cf1b83-8ac8-475a-bd69-9a80533ccbf7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.386503] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 713.386503] env[65385]: value = "task-4453426" [ 713.386503] env[65385]: _type = "Task" [ 713.386503] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.400590] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453426, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.431909] env[65385]: DEBUG nova.compute.utils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 713.431909] env[65385]: DEBUG nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 713.431909] env[65385]: DEBUG nova.network.neutron [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 713.432737] env[65385]: WARNING neutronclient.v2_0.client [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.433207] env[65385]: WARNING neutronclient.v2_0.client [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.433881] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.434368] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.468134] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3e0aec8-7f2b-4dbe-855c-d04363a819cf tempest-ImagesNegativeTestJSON-1486290601 tempest-ImagesNegativeTestJSON-1486290601-project-member] Lock "a71d2fa6-006c-4f76-8ae8-467e149297da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.102s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.487088] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "refresh_cache-90c36f67-0c25-4487-a5e4-89827290953b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.487293] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired lock "refresh_cache-90c36f67-0c25-4487-a5e4-89827290953b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.487533] env[65385]: DEBUG nova.network.neutron [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 713.561631] env[65385]: DEBUG nova.policy [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06826f37913b4dfe88fd709af43de5bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48e3431ecacb45288ccf0aba89cf5327', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 713.902033] env[65385]: DEBUG oslo_vmware.api [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453426, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262626} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.902210] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 713.902407] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 713.902486] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.902669] env[65385]: INFO nova.compute.manager [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Took 2.33 seconds to destroy the instance on the hypervisor. [ 713.902931] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 713.903217] env[65385]: DEBUG nova.compute.manager [-] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 713.904251] env[65385]: DEBUG nova.network.neutron [-] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 713.904251] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.904251] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.904687] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.938751] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 713.939021] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.945626] env[65385]: DEBUG nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 713.995305] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.995305] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.010069] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.063492] env[65385]: DEBUG nova.network.neutron [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 714.161825] env[65385]: DEBUG nova.network.neutron [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Successfully created port: 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 714.186140] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.186785] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.217065] env[65385]: DEBUG nova.network.neutron [-] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 714.551831] env[65385]: WARNING neutronclient.v2_0.client [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.552502] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.552856] env[65385]: WARNING openstack [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.688829] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Acquiring lock "1ad1e02b-fb82-4a88-9047-be316d590bd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.688930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.706875] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f3c4b2-ed62-4eb0-9664-6e6d5910d567 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.715854] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c3ba00-0aa6-4cd4-babb-2338f53ca08f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.719684] env[65385]: INFO nova.compute.manager [-] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Took 3.26 seconds to deallocate network for instance. [ 714.752051] env[65385]: DEBUG nova.network.neutron [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Updating instance_info_cache with network_info: [{"id": "c2cde5c4-0094-40e7-965c-60977c968a4d", "address": "fa:16:3e:42:cb:74", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cde5c4-00", "ovs_interfaceid": "c2cde5c4-0094-40e7-965c-60977c968a4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 714.756436] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27a2cda-ed8a-42d8-aeb2-0c6dc050a932 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.767240] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1fb7b44-1af9-4a28-9ef4-6ca535eb5984 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.787794] env[65385]: DEBUG nova.compute.provider_tree [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.821182] env[65385]: DEBUG nova.compute.manager [req-2cd9d950-53e7-4c33-b394-929682dc5c23 req-d0bcff67-ed9e-4dbb-aa96-3aaecea5fe62 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Received event network-vif-plugged-c2cde5c4-0094-40e7-965c-60977c968a4d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 714.821402] env[65385]: DEBUG oslo_concurrency.lockutils [req-2cd9d950-53e7-4c33-b394-929682dc5c23 req-d0bcff67-ed9e-4dbb-aa96-3aaecea5fe62 service nova] Acquiring lock "90c36f67-0c25-4487-a5e4-89827290953b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.822440] env[65385]: DEBUG oslo_concurrency.lockutils [req-2cd9d950-53e7-4c33-b394-929682dc5c23 req-d0bcff67-ed9e-4dbb-aa96-3aaecea5fe62 service nova] Lock "90c36f67-0c25-4487-a5e4-89827290953b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.822692] env[65385]: DEBUG oslo_concurrency.lockutils [req-2cd9d950-53e7-4c33-b394-929682dc5c23 req-d0bcff67-ed9e-4dbb-aa96-3aaecea5fe62 service nova] Lock "90c36f67-0c25-4487-a5e4-89827290953b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.822902] env[65385]: DEBUG nova.compute.manager [req-2cd9d950-53e7-4c33-b394-929682dc5c23 req-d0bcff67-ed9e-4dbb-aa96-3aaecea5fe62 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] No waiting events found dispatching network-vif-plugged-c2cde5c4-0094-40e7-965c-60977c968a4d {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 714.823142] env[65385]: WARNING nova.compute.manager [req-2cd9d950-53e7-4c33-b394-929682dc5c23 req-d0bcff67-ed9e-4dbb-aa96-3aaecea5fe62 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Received unexpected event network-vif-plugged-c2cde5c4-0094-40e7-965c-60977c968a4d for instance with vm_state building and task_state spawning. [ 714.942849] env[65385]: DEBUG nova.network.neutron [-] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 714.973242] env[65385]: DEBUG nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 714.990552] env[65385]: DEBUG nova.compute.manager [req-1ff1dd35-28f9-405e-a206-d6a917875460 req-4dd51796-f71e-4ca9-82db-61d1e108c65e service nova] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Received event network-vif-deleted-4343bbc8-b6b2-4198-9f4c-d6da5ff242dc {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 715.001430] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 715.001780] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 715.001979] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 715.002222] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 715.002385] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 715.002553] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 715.002792] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.002971] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 715.003257] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 715.003376] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 715.003557] env[65385]: DEBUG nova.virt.hardware [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 715.004834] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526b73b7-b0f2-431b-a89d-83de152c88d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.016964] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b1f484-78f0-4ee0-9349-0e35a2920611 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.261163] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Releasing lock "refresh_cache-90c36f67-0c25-4487-a5e4-89827290953b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.261422] env[65385]: DEBUG nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Instance network_info: |[{"id": "c2cde5c4-0094-40e7-965c-60977c968a4d", "address": "fa:16:3e:42:cb:74", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cde5c4-00", "ovs_interfaceid": "c2cde5c4-0094-40e7-965c-60977c968a4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 715.262493] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 715.263712] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:cb:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2cde5c4-0094-40e7-965c-60977c968a4d', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 715.274633] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Creating folder: Project (cb1ecd79d8e44f6096ca69a5238851b0). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 715.275686] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-880c7d5f-3884-4434-879d-208385b7c5f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.292023] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Created folder: Project (cb1ecd79d8e44f6096ca69a5238851b0) in parent group-v870881. [ 715.292377] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Creating folder: Instances. Parent ref: group-v870988. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 715.293498] env[65385]: DEBUG nova.scheduler.client.report [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 715.297162] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f9a85f03-7360-43a6-8d77-0469dfbecb06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.313089] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Created folder: Instances in parent group-v870988. [ 715.313588] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 715.313885] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 715.314354] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1800a2b-3e2b-46ea-a1f4-b69e2697e1f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.341045] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 715.341045] env[65385]: value = "task-4453429" [ 715.341045] env[65385]: _type = "Task" [ 715.341045] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.352156] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453429, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.449082] env[65385]: INFO nova.compute.manager [-] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Took 1.54 seconds to deallocate network for instance. [ 715.801843] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.883s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 715.802260] env[65385]: DEBUG nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 715.809229] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.519s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 715.809904] env[65385]: INFO nova.compute.claims [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.859882] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453429, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.034665] env[65385]: INFO nova.compute.manager [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Took 0.59 seconds to detach 1 volumes for instance. [ 716.059555] env[65385]: DEBUG nova.network.neutron [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Successfully updated port: 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 716.158438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "e3248f4a-8985-48b0-84cf-795f00ff21fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.158438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.189947] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.192166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.253704] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "f521e37d-aa86-4533-9fbe-7d040ed90e53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.254081] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.310797] env[65385]: DEBUG nova.compute.utils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 716.314530] env[65385]: DEBUG nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 716.314530] env[65385]: DEBUG nova.network.neutron [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 716.314530] env[65385]: WARNING neutronclient.v2_0.client [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.314792] env[65385]: WARNING neutronclient.v2_0.client [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.316207] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.316768] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.355195] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453429, 'name': CreateVM_Task, 'duration_secs': 0.629922} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.355379] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 716.356100] env[65385]: WARNING neutronclient.v2_0.client [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.356444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.356635] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.356984] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 716.357652] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cb275eb-e05d-4dbb-9289-5ace3d857da9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.364967] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 716.364967] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f30c0b-3897-c281-a44b-bf8c02fc2903" [ 716.364967] env[65385]: _type = "Task" [ 716.364967] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.376370] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f30c0b-3897-c281-a44b-bf8c02fc2903, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.447018] env[65385]: DEBUG nova.policy [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18c6274d42a04fbe81729aad5faf8a98', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cb1ecd79d8e44f6096ca69a5238851b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 716.548994] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.566870] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.566870] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquired lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.566870] env[65385]: DEBUG nova.network.neutron [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 716.814609] env[65385]: DEBUG nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 716.882768] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f30c0b-3897-c281-a44b-bf8c02fc2903, 'name': SearchDatastore_Task, 'duration_secs': 0.014947} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.882768] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.882768] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 716.883085] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.883131] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.883370] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.883583] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8a366de-8f4d-4914-a04a-9aac32064fea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.899647] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.899647] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 716.899647] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaf4c07c-921c-4534-ac7d-8c7a645dda29 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.911516] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 716.911516] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6ecac-7671-9e72-c38a-3e9036db628c" [ 716.911516] env[65385]: _type = "Task" [ 716.911516] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.914394] env[65385]: DEBUG nova.network.neutron [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Successfully created port: 828ba549-d28e-4ae9-bbf6-7050becb6e96 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 716.927473] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6ecac-7671-9e72-c38a-3e9036db628c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.070464] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.071102] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.226801] env[65385]: DEBUG nova.network.neutron [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 717.423029] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6ecac-7671-9e72-c38a-3e9036db628c, 'name': SearchDatastore_Task, 'duration_secs': 0.039963} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.427480] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19ddfb00-ae34-4351-8ef8-4fc0c5b4d966 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.435388] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 717.435388] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d438e9-b6cd-90ff-7035-17b7792564f8" [ 717.435388] env[65385]: _type = "Task" [ 717.435388] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.449272] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d438e9-b6cd-90ff-7035-17b7792564f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.462138] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15c6bc1-7d64-48b1-8765-8ed468c9963f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.471349] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a8d758-4b36-455c-8ce6-53850bb8a2c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.504859] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ca8559-4b28-4f4f-ae6b-e9db74b54030 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.514328] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7619ef48-2ce0-42d8-bbcd-986e9ef3bcd6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.530205] env[65385]: DEBUG nova.compute.provider_tree [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.708322] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52995cc4-34aa-1d5c-69ab-80111d1cbf79/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 717.709359] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b52774c-6fdc-4231-b311-738fa5a19711 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.719957] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52995cc4-34aa-1d5c-69ab-80111d1cbf79/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 717.720166] env[65385]: ERROR oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52995cc4-34aa-1d5c-69ab-80111d1cbf79/disk-0.vmdk due to incomplete transfer. [ 717.720421] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-92d6cf35-e32d-4dbb-bab8-d1d3e68c055a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.730379] env[65385]: DEBUG oslo_vmware.rw_handles [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52995cc4-34aa-1d5c-69ab-80111d1cbf79/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 717.730379] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Uploaded image 73727e37-3c18-44bd-8f2d-ba607b985a7a to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 717.733042] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 717.734523] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.734854] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.741432] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-005bb8eb-c789-4de9-b367-3032f976ac7d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.749864] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 717.749864] env[65385]: value = "task-4453430" [ 717.749864] env[65385]: _type = "Task" [ 717.749864] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.761540] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453430, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.825686] env[65385]: DEBUG nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 717.853914] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 717.854264] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 717.854458] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 717.854672] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 717.854819] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 717.854961] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 717.855191] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.855514] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 717.855747] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 717.855961] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 717.856308] env[65385]: DEBUG nova.virt.hardware [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 717.857724] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a215432-ca13-4aca-b163-968797676f97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.868598] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4fb423-11c3-4aff-8609-06d68d342e3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.947278] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d438e9-b6cd-90ff-7035-17b7792564f8, 'name': SearchDatastore_Task, 'duration_secs': 0.015803} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.948770] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 717.950042] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 90c36f67-0c25-4487-a5e4-89827290953b/90c36f67-0c25-4487-a5e4-89827290953b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 717.951495] env[65385]: DEBUG nova.compute.manager [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Received event network-changed-c2cde5c4-0094-40e7-965c-60977c968a4d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 717.951572] env[65385]: DEBUG nova.compute.manager [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Refreshing instance network info cache due to event network-changed-c2cde5c4-0094-40e7-965c-60977c968a4d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 717.951917] env[65385]: DEBUG oslo_concurrency.lockutils [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Acquiring lock "refresh_cache-90c36f67-0c25-4487-a5e4-89827290953b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.951917] env[65385]: DEBUG oslo_concurrency.lockutils [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Acquired lock "refresh_cache-90c36f67-0c25-4487-a5e4-89827290953b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.952078] env[65385]: DEBUG nova.network.neutron [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Refreshing network info cache for port c2cde5c4-0094-40e7-965c-60977c968a4d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 717.954386] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-832e6c35-0390-43a2-b04e-616a4c871a65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.964905] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 717.964905] env[65385]: value = "task-4453431" [ 717.964905] env[65385]: _type = "Task" [ 717.964905] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.979052] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.034372] env[65385]: DEBUG nova.scheduler.client.report [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 718.251764] env[65385]: WARNING neutronclient.v2_0.client [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.252523] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.252964] env[65385]: WARNING openstack [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.276709] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453430, 'name': Destroy_Task, 'duration_secs': 0.423082} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.277348] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Destroyed the VM [ 718.277541] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 718.277999] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2325f04b-4e16-4c29-aa5e-2040f6910d74 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.290271] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 718.290271] env[65385]: value = "task-4453432" [ 718.290271] env[65385]: _type = "Task" [ 718.290271] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.302952] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453432, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.455608] env[65385]: WARNING neutronclient.v2_0.client [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.456716] env[65385]: WARNING openstack [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.457182] env[65385]: WARNING openstack [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.475580] env[65385]: DEBUG nova.network.neutron [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 718.486765] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453431, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.540378] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.732s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.540697] env[65385]: DEBUG nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 718.544142] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.952s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.544142] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.549817] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.948s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.550036] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.554087] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.173s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.554339] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.556996] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.530s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.559219] env[65385]: INFO nova.compute.claims [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.569036] env[65385]: DEBUG nova.compute.manager [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Received event network-vif-plugged-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 718.569036] env[65385]: DEBUG oslo_concurrency.lockutils [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Acquiring lock "2f66a03b-874b-4bee-9694-49813ecb0c8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.569185] env[65385]: DEBUG oslo_concurrency.lockutils [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.569327] env[65385]: DEBUG oslo_concurrency.lockutils [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.569477] env[65385]: DEBUG nova.compute.manager [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] No waiting events found dispatching network-vif-plugged-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 718.569628] env[65385]: WARNING nova.compute.manager [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Received unexpected event network-vif-plugged-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 for instance with vm_state building and task_state spawning. [ 718.569780] env[65385]: DEBUG nova.compute.manager [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Received event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 718.569921] env[65385]: DEBUG nova.compute.manager [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing instance network info cache due to event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 718.570113] env[65385]: DEBUG oslo_concurrency.lockutils [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Acquiring lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.606953] env[65385]: INFO nova.scheduler.client.report [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Deleted allocations for instance 950efcc6-7e4f-4b30-b0fa-d940f893e1d5 [ 718.612020] env[65385]: INFO nova.scheduler.client.report [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleted allocations for instance 60067529-1071-4295-b1c9-21523bf347d9 [ 718.638373] env[65385]: INFO nova.scheduler.client.report [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Deleted allocations for instance 5291c04c-24ad-4c64-9fc2-fde42da4bc0f [ 718.695100] env[65385]: DEBUG nova.network.neutron [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Successfully updated port: 828ba549-d28e-4ae9-bbf6-7050becb6e96 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 718.777267] env[65385]: WARNING openstack [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.777267] env[65385]: WARNING openstack [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.806342] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453432, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.894555] env[65385]: WARNING neutronclient.v2_0.client [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.895268] env[65385]: WARNING openstack [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.895649] env[65385]: WARNING openstack [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.981807] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Releasing lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.982181] env[65385]: DEBUG nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Instance network_info: |[{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 718.982879] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544706} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.983259] env[65385]: DEBUG oslo_concurrency.lockutils [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Acquired lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.983513] env[65385]: DEBUG nova.network.neutron [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 718.985019] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:70:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b71230ae-e879-4384-88ce-fe64c86fce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.993532] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Creating folder: Project (48e3431ecacb45288ccf0aba89cf5327). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.993852] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 90c36f67-0c25-4487-a5e4-89827290953b/90c36f67-0c25-4487-a5e4-89827290953b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 718.994203] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.995620] env[65385]: WARNING neutronclient.v2_0.client [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.996287] env[65385]: WARNING openstack [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.996665] env[65385]: WARNING openstack [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.004294] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0738e3f7-13b1-4b54-9a73-1f9b499f1c7b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.006156] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b38ce95e-e0d2-44f6-b851-d388fcc64c37 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.011734] env[65385]: DEBUG nova.network.neutron [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Updated VIF entry in instance network info cache for port c2cde5c4-0094-40e7-965c-60977c968a4d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 719.012135] env[65385]: DEBUG nova.network.neutron [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Updating instance_info_cache with network_info: [{"id": "c2cde5c4-0094-40e7-965c-60977c968a4d", "address": "fa:16:3e:42:cb:74", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cde5c4-00", "ovs_interfaceid": "c2cde5c4-0094-40e7-965c-60977c968a4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 719.019642] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 719.019642] env[65385]: value = "task-4453434" [ 719.019642] env[65385]: _type = "Task" [ 719.019642] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.026025] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Created folder: Project (48e3431ecacb45288ccf0aba89cf5327) in parent group-v870881. [ 719.026293] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Creating folder: Instances. Parent ref: group-v870991. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 719.030101] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f2f4138-9215-4eb5-939f-6fa477508af8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.032376] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453434, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.043828] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Created folder: Instances in parent group-v870991. [ 719.044202] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 719.044469] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 719.044738] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0730b6ad-4ed4-4b40-9566-53d1b7077c03 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.064428] env[65385]: DEBUG nova.compute.utils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 719.066789] env[65385]: DEBUG nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 719.067043] env[65385]: DEBUG nova.network.neutron [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 719.067408] env[65385]: WARNING neutronclient.v2_0.client [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.067782] env[65385]: WARNING neutronclient.v2_0.client [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.068386] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.068800] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.082091] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 719.082091] env[65385]: value = "task-4453436" [ 719.082091] env[65385]: _type = "Task" [ 719.082091] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.092260] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453436, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.120607] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3adc1125-1822-42ee-a88c-72d95145fa2f tempest-ServerGroupTestJSON-1620569949 tempest-ServerGroupTestJSON-1620569949-project-member] Lock "950efcc6-7e4f-4b30-b0fa-d940f893e1d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.358s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 719.122373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b011aaca-da14-451d-8496-6fcdd12c1051 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "60067529-1071-4295-b1c9-21523bf347d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.089s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 719.134247] env[65385]: DEBUG nova.policy [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '039fe0672f174065ba37b3f9c57e5d41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a3e5ed6b95f4282b1ef29d17fa58975', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 719.145151] env[65385]: DEBUG oslo_concurrency.lockutils [None req-99b58f94-2775-49f9-8fe7-79279f210cdf tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "5291c04c-24ad-4c64-9fc2-fde42da4bc0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.728s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 719.150065] env[65385]: WARNING openstack [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.150687] env[65385]: WARNING openstack [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.203839] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "refresh_cache-483cf486-f0fc-48a4-9db2-970d0f35865a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.204083] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired lock "refresh_cache-483cf486-f0fc-48a4-9db2-970d0f35865a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.204355] env[65385]: DEBUG nova.network.neutron [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 719.302841] env[65385]: DEBUG oslo_vmware.api [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453432, 'name': RemoveSnapshot_Task, 'duration_secs': 0.546656} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.304015] env[65385]: WARNING neutronclient.v2_0.client [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.304452] env[65385]: WARNING openstack [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.304704] env[65385]: WARNING openstack [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.312631] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 719.312888] env[65385]: INFO nova.compute.manager [None req-aa54a8a9-9667-4389-85ff-462883f1b25b tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Took 14.58 seconds to snapshot the instance on the hypervisor. [ 719.469625] env[65385]: DEBUG nova.network.neutron [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updated VIF entry in instance network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 719.470305] env[65385]: DEBUG nova.network.neutron [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 719.516532] env[65385]: DEBUG oslo_concurrency.lockutils [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] Releasing lock "refresh_cache-90c36f67-0c25-4487-a5e4-89827290953b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.516647] env[65385]: DEBUG nova.compute.manager [req-d63b3039-40d7-44cf-b571-eb953743b86d req-d1907001-8b25-4bb6-972a-8965ef71a693 service nova] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Received event network-vif-deleted-9500dcc5-6478-4875-80fd-2ae8d2124100 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 719.537924] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453434, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145863} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.538291] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.539087] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34148d3-f57a-49bc-9fad-411c0256d155 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.566607] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] 90c36f67-0c25-4487-a5e4-89827290953b/90c36f67-0c25-4487-a5e4-89827290953b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.566607] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da8ce501-e38f-4b0c-a946-6c8033065418 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.585304] env[65385]: DEBUG nova.network.neutron [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Successfully created port: f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 719.587650] env[65385]: DEBUG nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 719.609269] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453436, 'name': CreateVM_Task, 'duration_secs': 0.470294} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.611677] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 719.612405] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 719.612405] env[65385]: value = "task-4453437" [ 719.612405] env[65385]: _type = "Task" [ 719.612405] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.612712] env[65385]: WARNING neutronclient.v2_0.client [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.613051] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.613245] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.613751] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 719.613927] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45f90082-e339-4116-9b73-182c13dd10f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.624021] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 719.624021] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528affa5-1951-6cb2-cbd8-26b63366a072" [ 719.624021] env[65385]: _type = "Task" [ 719.624021] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.628477] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453437, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.643950] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528affa5-1951-6cb2-cbd8-26b63366a072, 'name': SearchDatastore_Task, 'duration_secs': 0.011167} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.644045] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.644298] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.644534] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.644889] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.644889] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.645292] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1bff4882-9f4c-4811-97e5-e72f9fd33c76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.658581] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.658806] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 719.659684] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c839943d-b038-4b85-82c0-2e9659db9912 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.667176] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 719.667176] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d86c92-2a58-0237-2cef-48f04f606cef" [ 719.667176] env[65385]: _type = "Task" [ 719.667176] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.682429] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d86c92-2a58-0237-2cef-48f04f606cef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.708678] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.709164] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.973882] env[65385]: DEBUG oslo_concurrency.lockutils [req-b20769e6-3622-4447-a1ef-34db813b5cb2 req-ac16d429-169f-446a-921f-a8eae22776f3 service nova] Releasing lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.026602] env[65385]: DEBUG nova.network.neutron [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 720.128464] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453437, 'name': ReconfigVM_Task, 'duration_secs': 0.302085} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.128794] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Reconfigured VM instance instance-00000023 to attach disk [datastore1] 90c36f67-0c25-4487-a5e4-89827290953b/90c36f67-0c25-4487-a5e4-89827290953b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.129436] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b94c0ce8-327b-412f-9822-53fea923dd70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.137962] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 720.137962] env[65385]: value = "task-4453438" [ 720.137962] env[65385]: _type = "Task" [ 720.137962] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.149325] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453438, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.151195] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fefeb5-6390-4bd9-abe6-12eb43e6a878 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.160112] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afee0e5-b545-4e5a-bf8e-8bcfa7e65fda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.200339] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.200994] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.209168] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad84f84-0fcb-45a6-8942-bb2a1e054d1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.216781] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d86c92-2a58-0237-2cef-48f04f606cef, 'name': SearchDatastore_Task, 'duration_secs': 0.011215} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.218106] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e742c481-6117-466b-b586-acbeaae4e109 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.224622] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcba70d-26f8-451e-8fd8-477bbe745a90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.231333] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 720.231333] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522816bf-1490-4fad-731f-68ce6ced3536" [ 720.231333] env[65385]: _type = "Task" [ 720.231333] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.244645] env[65385]: DEBUG nova.compute.provider_tree [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.253070] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522816bf-1490-4fad-731f-68ce6ced3536, 'name': SearchDatastore_Task, 'duration_secs': 0.010516} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.253388] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.253648] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 2f66a03b-874b-4bee-9694-49813ecb0c8a/2f66a03b-874b-4bee-9694-49813ecb0c8a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 720.253927] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5943d58d-ca65-40f3-b8dc-140135e8188b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.263690] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 720.263690] env[65385]: value = "task-4453439" [ 720.263690] env[65385]: _type = "Task" [ 720.263690] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.274108] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453439, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.600500] env[65385]: DEBUG nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 720.633361] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 720.633657] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 720.633811] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 720.633987] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 720.634143] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 720.634286] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 720.634493] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.634693] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 720.634807] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 720.634962] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 720.635169] env[65385]: DEBUG nova.virt.hardware [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 720.636130] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2cf299-2094-4efb-80f3-8fd6edd02bbc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.652864] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e21647-98a8-4038-a11d-781be0ffb93f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.657569] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453438, 'name': Rename_Task, 'duration_secs': 0.220785} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.657872] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 720.658596] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95e4dd3d-73f3-46a9-b00b-4e82c4b06b20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.672151] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 720.672151] env[65385]: value = "task-4453440" [ 720.672151] env[65385]: _type = "Task" [ 720.672151] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.682526] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.748812] env[65385]: DEBUG nova.scheduler.client.report [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 720.775978] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453439, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.068427] env[65385]: WARNING neutronclient.v2_0.client [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.069147] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.069725] env[65385]: WARNING openstack [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.184700] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453440, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.196150] env[65385]: DEBUG nova.network.neutron [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Successfully updated port: f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 721.254858] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.698s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.255528] env[65385]: DEBUG nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 721.269761] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.060s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.269761] env[65385]: INFO nova.compute.claims [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.283536] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453439, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522758} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.283693] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 2f66a03b-874b-4bee-9694-49813ecb0c8a/2f66a03b-874b-4bee-9694-49813ecb0c8a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.283876] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.284435] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-702d842c-d88a-45dc-811c-5c3a33610e97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.295293] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 721.295293] env[65385]: value = "task-4453441" [ 721.295293] env[65385]: _type = "Task" [ 721.295293] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.308046] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453441, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.690744] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453440, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.696670] env[65385]: DEBUG nova.compute.manager [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Received event network-vif-plugged-828ba549-d28e-4ae9-bbf6-7050becb6e96 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 721.698295] env[65385]: DEBUG oslo_concurrency.lockutils [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Acquiring lock "483cf486-f0fc-48a4-9db2-970d0f35865a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.698295] env[65385]: DEBUG oslo_concurrency.lockutils [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.698295] env[65385]: DEBUG oslo_concurrency.lockutils [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.698295] env[65385]: DEBUG nova.compute.manager [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] No waiting events found dispatching network-vif-plugged-828ba549-d28e-4ae9-bbf6-7050becb6e96 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 721.698295] env[65385]: WARNING nova.compute.manager [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Received unexpected event network-vif-plugged-828ba549-d28e-4ae9-bbf6-7050becb6e96 for instance with vm_state building and task_state spawning. [ 721.698295] env[65385]: DEBUG nova.compute.manager [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Received event network-changed-828ba549-d28e-4ae9-bbf6-7050becb6e96 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 721.698295] env[65385]: DEBUG nova.compute.manager [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Refreshing instance network info cache due to event network-changed-828ba549-d28e-4ae9-bbf6-7050becb6e96. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 721.698295] env[65385]: DEBUG oslo_concurrency.lockutils [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Acquiring lock "refresh_cache-483cf486-f0fc-48a4-9db2-970d0f35865a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.701057] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.701057] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.701057] env[65385]: DEBUG nova.network.neutron [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 721.707420] env[65385]: DEBUG nova.network.neutron [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Updating instance_info_cache with network_info: [{"id": "828ba549-d28e-4ae9-bbf6-7050becb6e96", "address": "fa:16:3e:4d:1a:53", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap828ba549-d2", "ovs_interfaceid": "828ba549-d28e-4ae9-bbf6-7050becb6e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 721.774375] env[65385]: DEBUG nova.compute.utils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 721.778804] env[65385]: DEBUG nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 721.779185] env[65385]: DEBUG nova.network.neutron [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 721.779747] env[65385]: WARNING neutronclient.v2_0.client [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.780164] env[65385]: WARNING neutronclient.v2_0.client [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.780803] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.781265] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.809486] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453441, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07819} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.811088] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.811088] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7774a18b-9e5c-4d0a-b146-89812dd659f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.839984] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] 2f66a03b-874b-4bee-9694-49813ecb0c8a/2f66a03b-874b-4bee-9694-49813ecb0c8a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.840897] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc19f447-fa7c-4d85-92c4-1c1e91fed219 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.870161] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 721.870161] env[65385]: value = "task-4453442" [ 721.870161] env[65385]: _type = "Task" [ 721.870161] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.880138] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453442, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.898961] env[65385]: DEBUG nova.policy [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7e18058b77c4ccaa8afdf15e85930a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e94f2a868e2d4d31af54f2e5d8c479f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 722.188292] env[65385]: DEBUG oslo_vmware.api [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453440, 'name': PowerOnVM_Task, 'duration_secs': 1.159652} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.188842] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 722.193020] env[65385]: INFO nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Took 10.02 seconds to spawn the instance on the hypervisor. [ 722.193020] env[65385]: DEBUG nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 722.193020] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c92247-2366-4165-95f1-d499ef287cec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.209318] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.209705] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.217122] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Releasing lock "refresh_cache-483cf486-f0fc-48a4-9db2-970d0f35865a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.217493] env[65385]: DEBUG nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Instance network_info: |[{"id": "828ba549-d28e-4ae9-bbf6-7050becb6e96", "address": "fa:16:3e:4d:1a:53", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap828ba549-d2", "ovs_interfaceid": "828ba549-d28e-4ae9-bbf6-7050becb6e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 722.218282] env[65385]: DEBUG oslo_concurrency.lockutils [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Acquired lock "refresh_cache-483cf486-f0fc-48a4-9db2-970d0f35865a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.218282] env[65385]: DEBUG nova.network.neutron [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Refreshing network info cache for port 828ba549-d28e-4ae9-bbf6-7050becb6e96 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 722.219585] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:1a:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '828ba549-d28e-4ae9-bbf6-7050becb6e96', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.227366] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 722.227902] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 722.228146] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6b9bda1-afa7-483b-973d-3c2035fec26b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.247733] env[65385]: DEBUG nova.network.neutron [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Successfully created port: ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 722.252533] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.252533] env[65385]: value = "task-4453443" [ 722.252533] env[65385]: _type = "Task" [ 722.252533] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.262325] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453443, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.272408] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.272678] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 722.272881] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.273096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 722.273278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 722.275608] env[65385]: INFO nova.compute.manager [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Terminating instance [ 722.279839] env[65385]: DEBUG nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 722.381030] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453442, 'name': ReconfigVM_Task, 'duration_secs': 0.337686} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.381809] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Reconfigured VM instance instance-00000024 to attach disk [datastore1] 2f66a03b-874b-4bee-9694-49813ecb0c8a/2f66a03b-874b-4bee-9694-49813ecb0c8a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.382102] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6eaa2a0-211e-40c3-8083-a1d58c1696e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.394071] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 722.394071] env[65385]: value = "task-4453444" [ 722.394071] env[65385]: _type = "Task" [ 722.394071] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.405942] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453444, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.442421] env[65385]: DEBUG nova.network.neutron [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 722.713078] env[65385]: INFO nova.compute.manager [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Took 40.62 seconds to build instance. [ 722.728755] env[65385]: WARNING neutronclient.v2_0.client [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.729488] env[65385]: WARNING openstack [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.729839] env[65385]: WARNING openstack [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.765159] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453443, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.783260] env[65385]: DEBUG nova.compute.manager [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 722.783584] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.790928] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb91c40a-8d37-4cad-a230-ab6a189d69ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.801603] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 722.802801] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb093083-4ca6-4482-9376-b2665f28091c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.875855] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5591ea8f-7632-4134-9dc6-2cdfdd1ffd98 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.880209] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 722.880422] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 722.880599] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleting the datastore file [datastore2] 54ae4cd5-76d2-4e76-9528-6f511c4dca23 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 722.881336] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4c714dc-a2c4-4403-b157-a8fcafe5a323 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.887340] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6a1187-28ce-4a9d-8d2a-34698fa5305d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.891928] env[65385]: DEBUG oslo_vmware.api [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 722.891928] env[65385]: value = "task-4453446" [ 722.891928] env[65385]: _type = "Task" [ 722.891928] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.929917] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67ebe86-c636-4047-b17a-3f19e24cd61f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.936059] env[65385]: DEBUG oslo_vmware.api [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.941940] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453444, 'name': Rename_Task, 'duration_secs': 0.156024} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.944275] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 722.944653] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c128b95-46f2-4b77-b39b-ea0efa5b6664 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.947456] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b9d9fa-22f8-40f5-a364-8161dc6c9db5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.962140] env[65385]: DEBUG nova.compute.provider_tree [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.965572] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 722.965572] env[65385]: value = "task-4453447" [ 722.965572] env[65385]: _type = "Task" [ 722.965572] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.975921] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453447, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.126418] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.127024] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.215803] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dd221a-7515-4e11-b401-c3d3701fbe0a tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "90c36f67-0c25-4487-a5e4-89827290953b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.143s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.266768] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453443, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.298163] env[65385]: DEBUG nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 723.336794] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 723.337548] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 723.338205] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 723.338205] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 723.339257] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 723.339257] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 723.339257] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.339257] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 723.339891] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 723.340553] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 723.341054] env[65385]: DEBUG nova.virt.hardware [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 723.342297] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4910314-a3da-41ae-93d7-5190b35f393c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.357690] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc23ef5-aad7-4e90-b1a4-01836daf484e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.412609] env[65385]: DEBUG oslo_vmware.api [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201761} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.412928] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 723.412928] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 723.413074] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.413205] env[65385]: INFO nova.compute.manager [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Took 0.63 seconds to destroy the instance on the hypervisor. [ 723.413606] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 723.413694] env[65385]: DEBUG nova.compute.manager [-] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 723.413727] env[65385]: DEBUG nova.network.neutron [-] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 723.413960] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.414523] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.414872] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.460554] env[65385]: WARNING neutronclient.v2_0.client [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.462180] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.462180] env[65385]: WARNING openstack [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.470467] env[65385]: DEBUG nova.scheduler.client.report [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 723.487050] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453447, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.491764] env[65385]: WARNING openstack [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.492215] env[65385]: WARNING openstack [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.566528] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.718361] env[65385]: DEBUG nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 723.767416] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453443, 'name': CreateVM_Task, 'duration_secs': 1.507224} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.767702] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 723.768327] env[65385]: WARNING neutronclient.v2_0.client [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.768798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.768920] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.769302] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 723.770705] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3ca1fd1-47b1-4f04-8c32-ca44adcbae00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.775531] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 723.775531] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1eda3-d020-9251-b9f9-df2ee7747fc2" [ 723.775531] env[65385]: _type = "Task" [ 723.775531] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.787063] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1eda3-d020-9251-b9f9-df2ee7747fc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.982030] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.714s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.982283] env[65385]: DEBUG nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 723.991674] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.846s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.992694] env[65385]: DEBUG nova.objects.instance [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lazy-loading 'resources' on Instance uuid 8ac44124-d8cf-421c-9ae4-c943df94550a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 723.993251] env[65385]: DEBUG oslo_vmware.api [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453447, 'name': PowerOnVM_Task, 'duration_secs': 0.707251} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.996042] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 723.996359] env[65385]: INFO nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Took 9.02 seconds to spawn the instance on the hypervisor. [ 723.996568] env[65385]: DEBUG nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 723.997448] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea5fd02-3314-4cfd-af18-29cfdb1513e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.079799] env[65385]: DEBUG nova.network.neutron [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Successfully updated port: ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 724.128676] env[65385]: DEBUG nova.network.neutron [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updating instance_info_cache with network_info: [{"id": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "address": "fa:16:3e:01:88:e2", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf61ecab0-44", "ovs_interfaceid": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.222255] env[65385]: WARNING neutronclient.v2_0.client [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.222883] env[65385]: WARNING openstack [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.223254] env[65385]: WARNING openstack [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.250720] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 724.289638] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1eda3-d020-9251-b9f9-df2ee7747fc2, 'name': SearchDatastore_Task, 'duration_secs': 0.020753} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.290019] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.290783] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 724.290783] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.290783] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.290913] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 724.293060] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cac071f0-66d7-4ea1-9611-569e3e4de452 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.307355] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 724.307574] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 724.308398] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b74ac81-76b7-4be8-9d2d-78282b8f5f7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.318217] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 724.318217] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284a973-26ee-bc55-b625-a26cf4de6853" [ 724.318217] env[65385]: _type = "Task" [ 724.318217] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.330333] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284a973-26ee-bc55-b625-a26cf4de6853, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.499376] env[65385]: DEBUG nova.compute.utils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 724.500749] env[65385]: DEBUG nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 724.500938] env[65385]: DEBUG nova.network.neutron [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 724.505022] env[65385]: WARNING neutronclient.v2_0.client [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.505022] env[65385]: WARNING neutronclient.v2_0.client [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.505022] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.505022] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.525936] env[65385]: DEBUG nova.network.neutron [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Updated VIF entry in instance network info cache for port 828ba549-d28e-4ae9-bbf6-7050becb6e96. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 724.527440] env[65385]: DEBUG nova.network.neutron [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Updating instance_info_cache with network_info: [{"id": "828ba549-d28e-4ae9-bbf6-7050becb6e96", "address": "fa:16:3e:4d:1a:53", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap828ba549-d2", "ovs_interfaceid": "828ba549-d28e-4ae9-bbf6-7050becb6e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.529200] env[65385]: INFO nova.compute.manager [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Took 42.44 seconds to build instance. [ 724.588335] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.588570] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.589944] env[65385]: DEBUG nova.network.neutron [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 724.634397] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Releasing lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.635667] env[65385]: DEBUG nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Instance network_info: |[{"id": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "address": "fa:16:3e:01:88:e2", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf61ecab0-44", "ovs_interfaceid": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 724.637085] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:88:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '136c3499-9ca0-4f85-903d-1f194aa66ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f61ecab0-4449-47db-8bfb-6ff1499f6d71', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.644628] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Creating folder: Project (5a3e5ed6b95f4282b1ef29d17fa58975). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 724.651632] env[65385]: DEBUG nova.policy [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ba4783b423a4120a1035fabb02cd0aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9a1bf7c2dad4808a3782a3fba8a25fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 724.661028] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02cc1202-e3e1-4443-ba30-f055eb880f4d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.674810] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Created folder: Project (5a3e5ed6b95f4282b1ef29d17fa58975) in parent group-v870881. [ 724.675117] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Creating folder: Instances. Parent ref: group-v870995. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 724.675470] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e91af74f-f0f6-4f7a-ac54-587a5f70a89b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.688025] env[65385]: DEBUG nova.network.neutron [-] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.693194] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Created folder: Instances in parent group-v870995. [ 724.693568] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 724.696988] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 724.697265] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc7cc7e0-c520-42aa-99eb-b6935a846e73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.733225] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.733225] env[65385]: value = "task-4453450" [ 724.733225] env[65385]: _type = "Task" [ 724.733225] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.743765] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453450, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.830441] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284a973-26ee-bc55-b625-a26cf4de6853, 'name': SearchDatastore_Task, 'duration_secs': 0.019188} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.835525] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e0f056b-88b2-4c7d-95a2-fb0e2307940b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.842831] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 724.842831] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5240ac93-0464-a26a-9ce9-36395da11bbb" [ 724.842831] env[65385]: _type = "Task" [ 724.842831] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.859049] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5240ac93-0464-a26a-9ce9-36395da11bbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.012710] env[65385]: DEBUG nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 725.029317] env[65385]: DEBUG oslo_concurrency.lockutils [req-7094faa1-0d02-4f7f-b754-72160effca92 req-f594e1ce-4d1d-4268-99d8-729022830221 service nova] Releasing lock "refresh_cache-483cf486-f0fc-48a4-9db2-970d0f35865a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.031169] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7cf85a1d-327a-481f-a882-cecdf8cb6172 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.960s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.062920] env[65385]: DEBUG nova.network.neutron [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Successfully created port: 884a69db-a84a-46ae-9af6-bd5f21a386e7 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 725.095160] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.095160] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.199814] env[65385]: INFO nova.compute.manager [-] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Took 1.78 seconds to deallocate network for instance. [ 725.223893] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db1fe1c-5b9d-4350-aac6-b2fe285fc539 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.233690] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b98f4f-3dbc-4041-94ce-963e95cbf30f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.271588] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453450, 'name': CreateVM_Task, 'duration_secs': 0.391542} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.275022] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 725.275022] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a67e8d0-7afc-4bc3-93ba-a3051151c1a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.276340] env[65385]: WARNING neutronclient.v2_0.client [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.276847] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.277584] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.278033] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 725.278712] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c59e1ba-b532-418e-bff8-8f0f70160ea5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.293919] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 725.293919] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528f1360-9bf6-984e-624a-fe858a523471" [ 725.293919] env[65385]: _type = "Task" [ 725.293919] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.294695] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3dca593-64b3-47c4-8811-daa10964c43f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.323917] env[65385]: DEBUG nova.compute.provider_tree [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 725.327355] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528f1360-9bf6-984e-624a-fe858a523471, 'name': SearchDatastore_Task, 'duration_secs': 0.01496} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.327355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.327355] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.327355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.327355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.327355] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.327625] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00920ca0-f713-40a6-99d1-05f9b54d5fa3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.339369] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.339565] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 725.340681] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-130237f6-4325-484d-9fe8-1ead3a145421 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.352835] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 725.352835] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aea6a3-31ef-ed56-26db-d59840640480" [ 725.352835] env[65385]: _type = "Task" [ 725.352835] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.356511] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5240ac93-0464-a26a-9ce9-36395da11bbb, 'name': SearchDatastore_Task, 'duration_secs': 0.0163} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.362109] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.362109] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 483cf486-f0fc-48a4-9db2-970d0f35865a/483cf486-f0fc-48a4-9db2-970d0f35865a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 725.362356] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d9081e1-f404-4f65-b84b-efe131e74a64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.374339] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aea6a3-31ef-ed56-26db-d59840640480, 'name': SearchDatastore_Task, 'duration_secs': 0.012651} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.377307] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 725.377307] env[65385]: value = "task-4453451" [ 725.377307] env[65385]: _type = "Task" [ 725.377307] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.377547] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93025169-bef0-49b6-af47-4684f3f5a083 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.391751] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 725.391751] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520deebb-2ce0-fd6b-1463-3ee3f5a1abd6" [ 725.391751] env[65385]: _type = "Task" [ 725.391751] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.392210] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453451, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.402452] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520deebb-2ce0-fd6b-1463-3ee3f5a1abd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.408927] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "98430d7e-626e-4532-802e-38f7fc583ae3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.408927] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "98430d7e-626e-4532-802e-38f7fc583ae3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.415868] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "c5d00bb5-d869-466f-826e-61622e57959f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.416394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "c5d00bb5-d869-466f-826e-61622e57959f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.535312] env[65385]: DEBUG nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 725.539795] env[65385]: DEBUG nova.network.neutron [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 725.704631] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.847224] env[65385]: ERROR nova.scheduler.client.report [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] [req-0e4a70d9-1d9a-4ad1-884c-3ff3cb40fa27] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0e4a70d9-1d9a-4ad1-884c-3ff3cb40fa27"}]} [ 725.864808] env[65385]: DEBUG nova.scheduler.client.report [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 725.883364] env[65385]: DEBUG nova.scheduler.client.report [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 725.883692] env[65385]: DEBUG nova.compute.provider_tree [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 725.900111] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453451, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.901195] env[65385]: DEBUG nova.scheduler.client.report [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 725.910586] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520deebb-2ce0-fd6b-1463-3ee3f5a1abd6, 'name': SearchDatastore_Task, 'duration_secs': 0.013972} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.910888] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.911165] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 7fcbdc79-688c-479f-94e2-f4542abe714b/7fcbdc79-688c-479f-94e2-f4542abe714b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 725.911739] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-048a7204-4369-48ca-b86d-501dbce73851 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.920647] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 725.920647] env[65385]: value = "task-4453452" [ 725.920647] env[65385]: _type = "Task" [ 725.920647] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.927315] env[65385]: DEBUG nova.scheduler.client.report [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 725.936919] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.024486] env[65385]: DEBUG nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 726.065577] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 726.065988] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 726.066230] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 726.066909] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 726.066909] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 726.066909] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 726.067079] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.067163] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 726.067331] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 726.067500] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 726.067722] env[65385]: DEBUG nova.virt.hardware [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 726.068786] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adc606e-39bc-4dce-b968-475ecc418904 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.073110] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.083172] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ffab25-21bb-4d96-a596-a0006e389ae3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.399220] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453451, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.915511} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.400870] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 483cf486-f0fc-48a4-9db2-970d0f35865a/483cf486-f0fc-48a4-9db2-970d0f35865a.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 726.401533] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.402282] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0806cda4-4f81-4d75-a3d3-6f18766df350 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.413051] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 726.413051] env[65385]: value = "task-4453453" [ 726.413051] env[65385]: _type = "Task" [ 726.413051] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.428512] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.437229] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453452, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.442042] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.442293] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.532520] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921af313-5f94-42c7-9aa9-4cac6e90e3e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.543657] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5142f83f-6e54-4a33-b349-433150923239 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.589565] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84585df7-b3e1-4c9e-ba57-199cb71c6e97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.604204] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69c176b-401a-4252-ad40-809bdcfb7a1a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.622430] env[65385]: DEBUG nova.compute.provider_tree [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 726.783956] env[65385]: WARNING neutronclient.v2_0.client [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.785657] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.785766] env[65385]: WARNING openstack [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.820231] env[65385]: DEBUG nova.network.neutron [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Successfully updated port: 884a69db-a84a-46ae-9af6-bd5f21a386e7 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 726.920400] env[65385]: DEBUG nova.network.neutron [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updating instance_info_cache with network_info: [{"id": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "address": "fa:16:3e:a0:d3:7d", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff2751e9-ab", "ovs_interfaceid": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 726.930643] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.280097} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.931456] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 726.932693] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e9878f-3a39-4c00-b2e9-1250b38924ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.938813] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.848079} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.944025] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 7fcbdc79-688c-479f-94e2-f4542abe714b/7fcbdc79-688c-479f-94e2-f4542abe714b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 726.944025] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.944025] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2cdbb37-3e0c-4d4f-bba3-2c2981c6d5b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.964274] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 483cf486-f0fc-48a4-9db2-970d0f35865a/483cf486-f0fc-48a4-9db2-970d0f35865a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 726.969708] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e619120-39df-4e58-a53f-8107e4d0a884 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.988930] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 726.988930] env[65385]: value = "task-4453454" [ 726.988930] env[65385]: _type = "Task" [ 726.988930] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.997033] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 726.997033] env[65385]: value = "task-4453455" [ 726.997033] env[65385]: _type = "Task" [ 726.997033] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.000507] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453454, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.012655] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453455, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.168420] env[65385]: DEBUG nova.scheduler.client.report [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 66 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 727.168696] env[65385]: DEBUG nova.compute.provider_tree [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 66 to 67 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 727.168870] env[65385]: DEBUG nova.compute.provider_tree [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 727.208574] env[65385]: DEBUG nova.compute.manager [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received event network-vif-plugged-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 727.208847] env[65385]: DEBUG oslo_concurrency.lockutils [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Acquiring lock "7fcbdc79-688c-479f-94e2-f4542abe714b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.209069] env[65385]: DEBUG oslo_concurrency.lockutils [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.209237] env[65385]: DEBUG oslo_concurrency.lockutils [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.209403] env[65385]: DEBUG nova.compute.manager [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] No waiting events found dispatching network-vif-plugged-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 727.209665] env[65385]: WARNING nova.compute.manager [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received unexpected event network-vif-plugged-f61ecab0-4449-47db-8bfb-6ff1499f6d71 for instance with vm_state building and task_state spawning. [ 727.209823] env[65385]: DEBUG nova.compute.manager [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 727.209973] env[65385]: DEBUG nova.compute.manager [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing instance network info cache due to event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 727.210196] env[65385]: DEBUG oslo_concurrency.lockutils [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Acquiring lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.210364] env[65385]: DEBUG oslo_concurrency.lockutils [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Acquired lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.210527] env[65385]: DEBUG nova.network.neutron [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 727.235957] env[65385]: DEBUG nova.compute.manager [req-061ee29a-5ee5-4bbd-b27e-52d881a0e716 req-a5d61d5a-9dc7-4ff9-8377-fa0d07323891 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Received event network-vif-plugged-ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 727.236413] env[65385]: DEBUG oslo_concurrency.lockutils [req-061ee29a-5ee5-4bbd-b27e-52d881a0e716 req-a5d61d5a-9dc7-4ff9-8377-fa0d07323891 service nova] Acquiring lock "c798006c-3e74-461f-8397-3d8ca0be0a95-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.236413] env[65385]: DEBUG oslo_concurrency.lockutils [req-061ee29a-5ee5-4bbd-b27e-52d881a0e716 req-a5d61d5a-9dc7-4ff9-8377-fa0d07323891 service nova] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.236701] env[65385]: DEBUG oslo_concurrency.lockutils [req-061ee29a-5ee5-4bbd-b27e-52d881a0e716 req-a5d61d5a-9dc7-4ff9-8377-fa0d07323891 service nova] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.236752] env[65385]: DEBUG nova.compute.manager [req-061ee29a-5ee5-4bbd-b27e-52d881a0e716 req-a5d61d5a-9dc7-4ff9-8377-fa0d07323891 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] No waiting events found dispatching network-vif-plugged-ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 727.236969] env[65385]: WARNING nova.compute.manager [req-061ee29a-5ee5-4bbd-b27e-52d881a0e716 req-a5d61d5a-9dc7-4ff9-8377-fa0d07323891 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Received unexpected event network-vif-plugged-ff2751e9-ab94-4f90-84f6-0d47f1514411 for instance with vm_state building and task_state spawning. [ 727.328529] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.328722] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.328896] env[65385]: DEBUG nova.network.neutron [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 727.428286] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Releasing lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.428691] env[65385]: DEBUG nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Instance network_info: |[{"id": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "address": "fa:16:3e:a0:d3:7d", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff2751e9-ab", "ovs_interfaceid": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 727.429188] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:d3:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff2751e9-ab94-4f90-84f6-0d47f1514411', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.438557] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 727.438817] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 727.439066] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61dd08d5-c5a0-4338-bbe9-30ad4b1bc3bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.464745] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.464745] env[65385]: value = "task-4453456" [ 727.464745] env[65385]: _type = "Task" [ 727.464745] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.476393] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453456, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.504170] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453454, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.33267} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.505127] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 727.515724] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b248ef4-28b3-4394-96fb-0733c1a008ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.526235] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453455, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.547898] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 7fcbdc79-688c-479f-94e2-f4542abe714b/7fcbdc79-688c-479f-94e2-f4542abe714b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 727.548289] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54daa7ee-77e9-45d6-a745-99d3d84e5d50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.573041] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 727.573041] env[65385]: value = "task-4453457" [ 727.573041] env[65385]: _type = "Task" [ 727.573041] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.584474] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453457, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.677114] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.685s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.679543] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.169s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.681160] env[65385]: INFO nova.compute.claims [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.698037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.698323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.715181] env[65385]: INFO nova.scheduler.client.report [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Deleted allocations for instance 8ac44124-d8cf-421c-9ae4-c943df94550a [ 727.715181] env[65385]: WARNING neutronclient.v2_0.client [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.715181] env[65385]: WARNING openstack [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.715708] env[65385]: WARNING openstack [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.832878] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.833014] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.846364] env[65385]: WARNING openstack [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.846748] env[65385]: WARNING openstack [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.885095] env[65385]: DEBUG nova.network.neutron [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 727.907926] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.908406] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.941097] env[65385]: WARNING neutronclient.v2_0.client [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.941765] env[65385]: WARNING openstack [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.942135] env[65385]: WARNING openstack [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.983322] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453456, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.019203] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453455, 'name': ReconfigVM_Task, 'duration_secs': 0.994202} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.019449] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 483cf486-f0fc-48a4-9db2-970d0f35865a/483cf486-f0fc-48a4-9db2-970d0f35865a.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 728.020158] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-892d047e-6e47-4c88-8316-fb81cb14c9ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.028117] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 728.028117] env[65385]: value = "task-4453458" [ 728.028117] env[65385]: _type = "Task" [ 728.028117] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.035011] env[65385]: WARNING neutronclient.v2_0.client [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.035761] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.036122] env[65385]: WARNING openstack [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.047498] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453458, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.072145] env[65385]: DEBUG nova.network.neutron [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updated VIF entry in instance network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 728.072701] env[65385]: DEBUG nova.network.neutron [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updating instance_info_cache with network_info: [{"id": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "address": "fa:16:3e:01:88:e2", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf61ecab0-44", "ovs_interfaceid": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 728.083805] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453457, 'name': ReconfigVM_Task, 'duration_secs': 0.371975} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.084109] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 7fcbdc79-688c-479f-94e2-f4542abe714b/7fcbdc79-688c-479f-94e2-f4542abe714b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 728.085987] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04e0fa52-774d-43e4-9cdc-0a615bd0bcd6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.095675] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 728.095675] env[65385]: value = "task-4453459" [ 728.095675] env[65385]: _type = "Task" [ 728.095675] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.105683] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453459, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.235154] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ecb77469-f619-49ab-9365-704e8d124e75 tempest-VolumesAssistedSnapshotsTest-1436650459 tempest-VolumesAssistedSnapshotsTest-1436650459-project-member] Lock "8ac44124-d8cf-421c-9ae4-c943df94550a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.658s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.257216] env[65385]: DEBUG nova.network.neutron [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Updating instance_info_cache with network_info: [{"id": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "address": "fa:16:3e:91:29:10", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap884a69db-a8", "ovs_interfaceid": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 728.477684] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453456, 'name': CreateVM_Task, 'duration_secs': 0.568454} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.477684] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 728.478377] env[65385]: WARNING neutronclient.v2_0.client [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.478533] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.478701] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.479052] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 728.479320] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61ff7cbe-0d9b-421f-adb8-e0d0e3b31f0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.485385] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 728.485385] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6a364-ae73-01d2-dd52-de9d38fde1c1" [ 728.485385] env[65385]: _type = "Task" [ 728.485385] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.496286] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6a364-ae73-01d2-dd52-de9d38fde1c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.542022] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453458, 'name': Rename_Task, 'duration_secs': 0.236666} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.542022] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 728.542022] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1aedcd7-af75-4c46-b2b9-f2637d69f7db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.549536] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 728.549536] env[65385]: value = "task-4453460" [ 728.549536] env[65385]: _type = "Task" [ 728.549536] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.559460] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453460, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.579592] env[65385]: DEBUG oslo_concurrency.lockutils [req-c7df3e9d-5069-454e-a0aa-9f03aab986f9 req-3563540e-8670-4280-9e6c-44f1371b7395 service nova] Releasing lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.609015] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453459, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.761052] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.761966] env[65385]: DEBUG nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Instance network_info: |[{"id": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "address": "fa:16:3e:91:29:10", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap884a69db-a8", "ovs_interfaceid": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 728.762131] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:29:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6815237d-f565-474d-a3c0-9c675478eb00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '884a69db-a84a-46ae-9af6-bd5f21a386e7', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 728.771744] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 728.775639] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 728.776172] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d30a90d7-67ba-4bf2-8481-5048544aa13a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.810497] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 728.810497] env[65385]: value = "task-4453461" [ 728.810497] env[65385]: _type = "Task" [ 728.810497] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.820780] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453461, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.005857] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "ab8df1d8-615d-4083-bdf4-25f679a64982" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.006338] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.006659] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6a364-ae73-01d2-dd52-de9d38fde1c1, 'name': SearchDatastore_Task, 'duration_secs': 0.022408} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.007208] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.007485] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.007670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.007827] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.008092] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.008882] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93eb7226-ac26-41cb-bdb9-aaa15191e157 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.025794] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.025971] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 729.026839] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fae785b-1e29-4975-a885-d32abc52cc08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.034314] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 729.034314] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523e6bcf-a16c-8935-2670-88b33b6f0e39" [ 729.034314] env[65385]: _type = "Task" [ 729.034314] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.049164] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523e6bcf-a16c-8935-2670-88b33b6f0e39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.059959] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453460, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.112496] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453459, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.325359] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453461, 'name': CreateVM_Task, 'duration_secs': 0.439845} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.325359] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 729.325359] env[65385]: WARNING neutronclient.v2_0.client [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.325359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.325359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.325359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 729.325359] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-919bd77a-8d01-4ea2-910f-f67670874414 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.331442] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 729.331442] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b36797-3892-d39e-b65d-85d894422a9d" [ 729.331442] env[65385]: _type = "Task" [ 729.331442] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.352051] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b36797-3892-d39e-b65d-85d894422a9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.429551] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731f258d-3a0f-44bf-8722-5a8362a5e7e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.439948] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057df380-1ae3-4e3a-bf4e-96bb3c3fcefe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.477444] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4fd350-64a4-44c4-a3e4-7ddaafeb5f4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.487477] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6552ec-cbaf-429e-9658-f5c3aa8a6fdd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.503750] env[65385]: DEBUG nova.compute.provider_tree [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 729.545517] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523e6bcf-a16c-8935-2670-88b33b6f0e39, 'name': SearchDatastore_Task, 'duration_secs': 0.01561} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.546369] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53a2dc22-b1c5-4db2-9f85-4ae5914dcf48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.557501] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 729.557501] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ce2f26-3e1d-d82d-1c6d-be1708fa3663" [ 729.557501] env[65385]: _type = "Task" [ 729.557501] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.558998] env[65385]: DEBUG nova.compute.manager [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Received event network-vif-deleted-6b67e7c9-05f5-47d9-9b81-86464d52033c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 729.559329] env[65385]: DEBUG nova.compute.manager [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Received event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 729.559518] env[65385]: DEBUG nova.compute.manager [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing instance network info cache due to event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 729.559734] env[65385]: DEBUG oslo_concurrency.lockutils [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Acquiring lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.559969] env[65385]: DEBUG oslo_concurrency.lockutils [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Acquired lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.560049] env[65385]: DEBUG nova.network.neutron [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 729.569831] env[65385]: DEBUG oslo_vmware.api [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453460, 'name': PowerOnVM_Task, 'duration_secs': 0.715047} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.569974] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 729.570215] env[65385]: INFO nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Took 11.74 seconds to spawn the instance on the hypervisor. [ 729.570565] env[65385]: DEBUG nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 729.571268] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d94e33-84e5-4fa1-be09-cee3ca6f9619 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.579312] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ce2f26-3e1d-d82d-1c6d-be1708fa3663, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.582717] env[65385]: DEBUG nova.compute.manager [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Received event network-changed-ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 729.583130] env[65385]: DEBUG nova.compute.manager [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Refreshing instance network info cache due to event network-changed-ff2751e9-ab94-4f90-84f6-0d47f1514411. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 729.583351] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Acquiring lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.583484] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Acquired lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.583630] env[65385]: DEBUG nova.network.neutron [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Refreshing network info cache for port ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 729.613486] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453459, 'name': Rename_Task, 'duration_secs': 1.211462} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.613866] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 729.615325] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60178920-bd70-4f8b-b283-2826bc0102a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.625024] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 729.625024] env[65385]: value = "task-4453462" [ 729.625024] env[65385]: _type = "Task" [ 729.625024] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.642565] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.845576] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b36797-3892-d39e-b65d-85d894422a9d, 'name': SearchDatastore_Task, 'duration_secs': 0.015635} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.845991] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.846311] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.846609] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.029879] env[65385]: ERROR nova.scheduler.client.report [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [req-756ad5dc-330b-48f4-ac16-73c316db81bc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-756ad5dc-330b-48f4-ac16-73c316db81bc"}]} [ 730.054212] env[65385]: DEBUG nova.scheduler.client.report [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 730.066839] env[65385]: WARNING neutronclient.v2_0.client [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.067597] env[65385]: WARNING openstack [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.067985] env[65385]: WARNING openstack [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.076348] env[65385]: DEBUG nova.scheduler.client.report [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 730.076551] env[65385]: DEBUG nova.compute.provider_tree [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 730.085419] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ce2f26-3e1d-d82d-1c6d-be1708fa3663, 'name': SearchDatastore_Task, 'duration_secs': 0.026961} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.085697] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.085952] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c798006c-3e74-461f-8397-3d8ca0be0a95/c798006c-3e74-461f-8397-3d8ca0be0a95.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 730.086247] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.086429] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 730.086679] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9209c4d3-650a-4bfd-9e49-c2f010cdec3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.089674] env[65385]: DEBUG nova.scheduler.client.report [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 730.091687] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69d66ee0-c6f6-40a5-8b94-b40a547cae68 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.094234] env[65385]: WARNING neutronclient.v2_0.client [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.094890] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.095432] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.114230] env[65385]: INFO nova.compute.manager [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Took 44.90 seconds to build instance. [ 730.120975] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 730.120975] env[65385]: value = "task-4453463" [ 730.120975] env[65385]: _type = "Task" [ 730.120975] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.130790] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 730.133818] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 730.133818] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12875057-a585-4099-a5b4-9b9ac51b6621 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.141703] env[65385]: DEBUG nova.scheduler.client.report [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 730.153040] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453463, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.154790] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 730.154790] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52165d88-172a-b7ac-d1a3-9dc7b83bc5db" [ 730.154790] env[65385]: _type = "Task" [ 730.154790] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.160971] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453462, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.173299] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52165d88-172a-b7ac-d1a3-9dc7b83bc5db, 'name': SearchDatastore_Task, 'duration_secs': 0.012251} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.174287] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f2f754a-0875-4cc7-98ab-a84132a41d6d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.182148] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 730.182148] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529f2754-68b6-a3e9-f884-7e34d17a5676" [ 730.182148] env[65385]: _type = "Task" [ 730.182148] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.197602] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529f2754-68b6-a3e9-f884-7e34d17a5676, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.623152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a42bd6e-b6eb-472c-b0aa-86490bc5df6b tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.419s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.644212] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453463, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.649823] env[65385]: DEBUG oslo_vmware.api [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453462, 'name': PowerOnVM_Task, 'duration_secs': 0.835593} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.655317] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 730.655317] env[65385]: INFO nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Took 10.05 seconds to spawn the instance on the hypervisor. [ 730.655495] env[65385]: DEBUG nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 730.657375] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9670c36-ac0c-4c6c-ae22-f403b2977938 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.707033] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529f2754-68b6-a3e9-f884-7e34d17a5676, 'name': SearchDatastore_Task, 'duration_secs': 0.01345} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.707033] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.707033] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 34730159-3bec-4ad5-b85e-0f67998d6001/34730159-3bec-4ad5-b85e-0f67998d6001.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 730.707033] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0541bbf6-9aa5-46a0-b7e6-4101c103a329 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.719248] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 730.719248] env[65385]: value = "task-4453464" [ 730.719248] env[65385]: _type = "Task" [ 730.719248] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.732262] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.776916] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7978860-f7e8-49ab-bf9d-4466dacbd9e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.785325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c8f982-cf65-4293-9f16-c93926ac64d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.834029] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf86ea4-343a-4c60-b175-c0b590d670af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.842067] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3528fbef-44fa-435b-9844-d2e6b0bff287 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.858411] env[65385]: DEBUG nova.compute.provider_tree [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 731.133618] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 731.140151] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453463, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613603} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.143771] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c798006c-3e74-461f-8397-3d8ca0be0a95/c798006c-3e74-461f-8397-3d8ca0be0a95.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 731.144758] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.146610] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09608d37-f854-4124-85cb-ccd338e6c041 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.158658] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 731.158658] env[65385]: value = "task-4453465" [ 731.158658] env[65385]: _type = "Task" [ 731.158658] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.169951] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.183965] env[65385]: INFO nova.compute.manager [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Took 45.95 seconds to build instance. [ 731.233106] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453464, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.411638] env[65385]: DEBUG nova.scheduler.client.report [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 69 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 731.411939] env[65385]: DEBUG nova.compute.provider_tree [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 69 to 70 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 731.412174] env[65385]: DEBUG nova.compute.provider_tree [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 731.676874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.682063] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197439} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.682063] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 731.682361] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3cdc67-331c-4ff0-8317-3534f1f9a611 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.687567] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bfd25ecf-a862-4fc0-be65-6ce9eed9b4cf tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.475s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.712247] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] c798006c-3e74-461f-8397-3d8ca0be0a95/c798006c-3e74-461f-8397-3d8ca0be0a95.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.712800] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd0f82d9-71e8-4186-9c37-5499796f2825 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.748147] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.976211} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.749679] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 34730159-3bec-4ad5-b85e-0f67998d6001/34730159-3bec-4ad5-b85e-0f67998d6001.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 731.749935] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.750275] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 731.750275] env[65385]: value = "task-4453466" [ 731.750275] env[65385]: _type = "Task" [ 731.750275] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.750555] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-212b6eca-0faa-43b8-a109-ff77ad5535de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.765315] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453466, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.767570] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 731.767570] env[65385]: value = "task-4453467" [ 731.767570] env[65385]: _type = "Task" [ 731.767570] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.779076] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453467, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.917618] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.238s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.918184] env[65385]: DEBUG nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 731.921339] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.788s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.921656] env[65385]: DEBUG nova.objects.instance [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 732.190822] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 732.265889] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.279249] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453467, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077863} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.281515] env[65385]: WARNING openstack [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.281882] env[65385]: WARNING openstack [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.290063] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.291222] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1404bd47-4705-48d9-b607-faebd185c829 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.317093] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 34730159-3bec-4ad5-b85e-0f67998d6001/34730159-3bec-4ad5-b85e-0f67998d6001.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.317467] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.317840] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.326400] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b9965a1-f5d4-4c88-badf-6e1e9d65e528 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.353577] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 732.353577] env[65385]: value = "task-4453468" [ 732.353577] env[65385]: _type = "Task" [ 732.353577] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.371303] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453468, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.435046] env[65385]: DEBUG nova.compute.utils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 732.435524] env[65385]: DEBUG nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 732.435864] env[65385]: DEBUG nova.network.neutron [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 732.436326] env[65385]: WARNING neutronclient.v2_0.client [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.436849] env[65385]: WARNING neutronclient.v2_0.client [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.437413] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.437766] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.719694] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.766151] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.775489] env[65385]: WARNING neutronclient.v2_0.client [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.775489] env[65385]: WARNING openstack [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.775489] env[65385]: WARNING openstack [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.807198] env[65385]: WARNING neutronclient.v2_0.client [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.807198] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.807198] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.869438] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453468, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.885038] env[65385]: DEBUG nova.policy [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae527c3a64084ef388ff9415f1a6f883', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99173b1f52aa42f9b0db92366db8c411', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 732.936738] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fee22c9-87dc-4ce8-b319-1a642c119ff1 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.940481] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.716s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.940481] env[65385]: INFO nova.compute.claims [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.948337] env[65385]: DEBUG nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 733.266011] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453466, 'name': ReconfigVM_Task, 'duration_secs': 1.356424} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.266388] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Reconfigured VM instance instance-00000027 to attach disk [datastore2] c798006c-3e74-461f-8397-3d8ca0be0a95/c798006c-3e74-461f-8397-3d8ca0be0a95.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.267646] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe6c82c9-f7ea-4f96-b066-e4afcd58d338 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.279197] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 733.279197] env[65385]: value = "task-4453469" [ 733.279197] env[65385]: _type = "Task" [ 733.279197] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.295818] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453469, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.326423] env[65385]: DEBUG nova.network.neutron [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Successfully created port: ad95d76f-e023-47b2-a936-e7ed11779276 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 733.367425] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453468, 'name': ReconfigVM_Task, 'duration_secs': 0.751403} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.368119] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 34730159-3bec-4ad5-b85e-0f67998d6001/34730159-3bec-4ad5-b85e-0f67998d6001.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.368695] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50acc70b-6599-4eea-8218-e02743d0d288 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.377680] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 733.377680] env[65385]: value = "task-4453470" [ 733.377680] env[65385]: _type = "Task" [ 733.377680] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.383902] env[65385]: DEBUG nova.network.neutron [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updated VIF entry in instance network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 733.384301] env[65385]: DEBUG nova.network.neutron [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 733.393902] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453470, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.550833] env[65385]: DEBUG nova.network.neutron [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updated VIF entry in instance network info cache for port ff2751e9-ab94-4f90-84f6-0d47f1514411. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 733.551914] env[65385]: DEBUG nova.network.neutron [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updating instance_info_cache with network_info: [{"id": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "address": "fa:16:3e:a0:d3:7d", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff2751e9-ab", "ovs_interfaceid": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 733.792213] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453469, 'name': Rename_Task, 'duration_secs': 0.198431} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.792766] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 733.793170] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2233bc7-28b1-4fb2-8bc3-f98a12474d4d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.803432] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 733.803432] env[65385]: value = "task-4453471" [ 733.803432] env[65385]: _type = "Task" [ 733.803432] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.814102] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.891841] env[65385]: DEBUG oslo_concurrency.lockutils [req-59d7a166-7d2a-483e-b21e-ad52648f7210 req-92d97c86-2572-45a4-8253-55889b55c49f service nova] Releasing lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 733.892276] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453470, 'name': Rename_Task, 'duration_secs': 0.239406} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.892670] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 733.892981] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81edc2c6-4b8e-4df0-b2c8-c3a61186dbb8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.901362] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 733.901362] env[65385]: value = "task-4453472" [ 733.901362] env[65385]: _type = "Task" [ 733.901362] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.911514] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453472, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.959264] env[65385]: DEBUG nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 733.996331] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:48:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f35b2975-b7d3-4f0c-9c41-d7d5c82d8413',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1230048441',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 733.996516] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 733.996664] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 733.996930] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 733.997504] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 733.997504] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 733.997714] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.997884] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 733.998072] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 733.998232] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 733.998395] env[65385]: DEBUG nova.virt.hardware [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 733.999654] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe99e53-c016-4cdb-af0d-f6f84cbdd7fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.014567] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e808290-73db-4f81-bc32-817ea904d014 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.056424] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Releasing lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 734.056424] env[65385]: DEBUG nova.compute.manager [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Received event network-vif-plugged-884a69db-a84a-46ae-9af6-bd5f21a386e7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 734.056424] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Acquiring lock "34730159-3bec-4ad5-b85e-0f67998d6001-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.056424] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Lock "34730159-3bec-4ad5-b85e-0f67998d6001-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.056626] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Lock "34730159-3bec-4ad5-b85e-0f67998d6001-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.056697] env[65385]: DEBUG nova.compute.manager [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] No waiting events found dispatching network-vif-plugged-884a69db-a84a-46ae-9af6-bd5f21a386e7 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 734.056865] env[65385]: WARNING nova.compute.manager [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Received unexpected event network-vif-plugged-884a69db-a84a-46ae-9af6-bd5f21a386e7 for instance with vm_state building and task_state spawning. [ 734.057482] env[65385]: DEBUG nova.compute.manager [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Received event network-changed-884a69db-a84a-46ae-9af6-bd5f21a386e7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 734.057482] env[65385]: DEBUG nova.compute.manager [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Refreshing instance network info cache due to event network-changed-884a69db-a84a-46ae-9af6-bd5f21a386e7. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 734.057482] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Acquiring lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.057714] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Acquired lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.057714] env[65385]: DEBUG nova.network.neutron [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Refreshing network info cache for port 884a69db-a84a-46ae-9af6-bd5f21a386e7 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 734.323294] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453471, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.419958] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453472, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.564191] env[65385]: WARNING neutronclient.v2_0.client [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.564370] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.564605] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.642630] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00549441-f5e4-4426-b013-276293433b6d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.651390] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3609bc82-fb18-453b-bf21-46959d6dade9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.686282] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494db609-1fb9-4bd6-99f1-76c8a02e7453 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.695397] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c27f795-db31-4c1c-a357-2aa504734c1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.710634] env[65385]: DEBUG nova.compute.provider_tree [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.815084] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453471, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.928044] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453472, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.159846] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.159846] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.216871] env[65385]: DEBUG nova.scheduler.client.report [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 735.319252] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453471, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.417244] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453472, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.562841] env[65385]: DEBUG nova.network.neutron [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Successfully updated port: ad95d76f-e023-47b2-a936-e7ed11779276 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 735.722429] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.784s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.722967] env[65385]: DEBUG nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 735.728677] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.769s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.728677] env[65385]: DEBUG nova.objects.instance [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lazy-loading 'resources' on Instance uuid 8b478626-e3dc-42c4-bd59-fa0a71d798c9 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 735.741656] env[65385]: WARNING neutronclient.v2_0.client [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.742109] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.742442] env[65385]: WARNING openstack [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.822913] env[65385]: DEBUG oslo_vmware.api [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453471, 'name': PowerOnVM_Task, 'duration_secs': 1.764365} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.823335] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 735.823407] env[65385]: INFO nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Took 12.53 seconds to spawn the instance on the hypervisor. [ 735.826980] env[65385]: DEBUG nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 735.826980] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56464021-b932-4adf-a06f-95daa0ff78f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.876349] env[65385]: DEBUG nova.network.neutron [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Updated VIF entry in instance network info cache for port 884a69db-a84a-46ae-9af6-bd5f21a386e7. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 735.877240] env[65385]: DEBUG nova.network.neutron [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Updating instance_info_cache with network_info: [{"id": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "address": "fa:16:3e:91:29:10", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap884a69db-a8", "ovs_interfaceid": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 735.919554] env[65385]: DEBUG oslo_vmware.api [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453472, 'name': PowerOnVM_Task, 'duration_secs': 1.707973} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.919554] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 735.919828] env[65385]: INFO nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Took 9.90 seconds to spawn the instance on the hypervisor. [ 735.919931] env[65385]: DEBUG nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 735.920835] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094c10f9-656e-4e2b-a079-219fa3cf84b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.069498] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.069498] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.069498] env[65385]: DEBUG nova.network.neutron [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 736.229218] env[65385]: DEBUG nova.compute.utils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 736.230426] env[65385]: DEBUG nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 736.361132] env[65385]: INFO nova.compute.manager [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Took 38.36 seconds to build instance. [ 736.380878] env[65385]: DEBUG oslo_concurrency.lockutils [req-56587499-077c-4b9e-b541-49ef92b24189 req-bdeac1ce-5dcc-49c5-bcd8-a9331b90954a service nova] Releasing lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.441984] env[65385]: INFO nova.compute.manager [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Took 36.26 seconds to build instance. [ 736.575065] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.575477] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 736.737608] env[65385]: DEBUG nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 736.862747] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e4e65bd5-9f59-46cb-83d7-6d3f61a440e6 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.874s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.927049] env[65385]: DEBUG nova.network.neutron [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 736.944849] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea16b28-1690-4122-9bbc-f93148880857 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.948775] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcad7967-7263-4246-827c-a49ca9fee060 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "34730159-3bec-4ad5-b85e-0f67998d6001" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.783s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.955916] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99bbc263-0601-4926-b056-950be80641cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.961653] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.962109] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 737.010907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6abb0f0-82cf-4e27-9d55-4f6bf0d37dcb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.018174] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6e07f9-3656-49b0-a46b-ae5072a99416 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.034912] env[65385]: DEBUG nova.compute.provider_tree [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.051533] env[65385]: WARNING neutronclient.v2_0.client [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 737.052560] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 737.052560] env[65385]: WARNING openstack [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 737.137050] env[65385]: DEBUG nova.compute.manager [req-c4351026-8d97-4db0-9ba4-368d207e97d3 req-3d5d46e6-6e43-4859-8673-b97c6384a92e service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Received event network-vif-plugged-ad95d76f-e023-47b2-a936-e7ed11779276 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 737.137244] env[65385]: DEBUG oslo_concurrency.lockutils [req-c4351026-8d97-4db0-9ba4-368d207e97d3 req-3d5d46e6-6e43-4859-8673-b97c6384a92e service nova] Acquiring lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.137564] env[65385]: DEBUG oslo_concurrency.lockutils [req-c4351026-8d97-4db0-9ba4-368d207e97d3 req-3d5d46e6-6e43-4859-8673-b97c6384a92e service nova] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 737.137960] env[65385]: DEBUG oslo_concurrency.lockutils [req-c4351026-8d97-4db0-9ba4-368d207e97d3 req-3d5d46e6-6e43-4859-8673-b97c6384a92e service nova] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 737.137960] env[65385]: DEBUG nova.compute.manager [req-c4351026-8d97-4db0-9ba4-368d207e97d3 req-3d5d46e6-6e43-4859-8673-b97c6384a92e service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] No waiting events found dispatching network-vif-plugged-ad95d76f-e023-47b2-a936-e7ed11779276 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 737.138194] env[65385]: WARNING nova.compute.manager [req-c4351026-8d97-4db0-9ba4-368d207e97d3 req-3d5d46e6-6e43-4859-8673-b97c6384a92e service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Received unexpected event network-vif-plugged-ad95d76f-e023-47b2-a936-e7ed11779276 for instance with vm_state building and task_state spawning. [ 737.163187] env[65385]: DEBUG nova.network.neutron [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance_info_cache with network_info: [{"id": "ad95d76f-e023-47b2-a936-e7ed11779276", "address": "fa:16:3e:b0:c7:d2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95d76f-e0", "ovs_interfaceid": "ad95d76f-e023-47b2-a936-e7ed11779276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 737.366137] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 737.453031] env[65385]: DEBUG nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 737.539297] env[65385]: DEBUG nova.scheduler.client.report [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 737.665433] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.665839] env[65385]: DEBUG nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Instance network_info: |[{"id": "ad95d76f-e023-47b2-a936-e7ed11779276", "address": "fa:16:3e:b0:c7:d2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95d76f-e0", "ovs_interfaceid": "ad95d76f-e023-47b2-a936-e7ed11779276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 737.666321] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:c7:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad95d76f-e023-47b2-a936-e7ed11779276', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.677114] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 737.679041] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 737.679041] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5c7092f-c884-4f02-87da-1ac017bc963a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.701251] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.701251] env[65385]: value = "task-4453473" [ 737.701251] env[65385]: _type = "Task" [ 737.701251] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.715387] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453473, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.751385] env[65385]: DEBUG nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 737.789030] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 737.789321] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 737.789501] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 737.789669] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 737.789749] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 737.789919] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 737.790532] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.790532] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 737.790532] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 737.791475] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 737.791475] env[65385]: DEBUG nova.virt.hardware [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 737.792276] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47251f8-483c-438d-9c3d-f560fa5f7344 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.802685] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c4ddf6-0644-495b-ad0e-9f00468838ae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.820036] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.826439] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Creating folder: Project (414acc709ba44c4d8d0e3b53a02ac787). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.826986] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-33d8c412-71f1-4701-ad60-16bb2e4f7a94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.842028] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Created folder: Project (414acc709ba44c4d8d0e3b53a02ac787) in parent group-v870881. [ 737.842028] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Creating folder: Instances. Parent ref: group-v871001. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.842028] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d42b2af-3a84-4461-82ee-fb909687db8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.849117] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "65f9bc55-7bab-43b0-a974-eb5080389b7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.849671] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "65f9bc55-7bab-43b0-a974-eb5080389b7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 737.859019] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Created folder: Instances in parent group-v871001. [ 737.859019] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 737.859019] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 737.859019] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a1cf91e-bf80-46bf-abc8-2a583397181e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.887253] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.887253] env[65385]: value = "task-4453476" [ 737.887253] env[65385]: _type = "Task" [ 737.887253] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.896936] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453476, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.899669] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.976363] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.047499] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.322s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.050454] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.002s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.050587] env[65385]: DEBUG nova.objects.instance [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lazy-loading 'resources' on Instance uuid 1bc94ee7-d4f9-48c3-97f1-9e662eb50582 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 738.080438] env[65385]: INFO nova.scheduler.client.report [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Deleted allocations for instance 8b478626-e3dc-42c4-bd59-fa0a71d798c9 [ 738.215733] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453473, 'name': CreateVM_Task, 'duration_secs': 0.505976} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.215943] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 738.216477] env[65385]: WARNING neutronclient.v2_0.client [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.217055] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.217235] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.217611] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 738.218064] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6583aff3-3189-430f-b4c2-6cffe1169a90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.224040] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 738.224040] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52903ed8-3bb3-7e2b-0d06-e48369e8e40d" [ 738.224040] env[65385]: _type = "Task" [ 738.224040] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.235422] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52903ed8-3bb3-7e2b-0d06-e48369e8e40d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.398205] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453476, 'name': CreateVM_Task, 'duration_secs': 0.444663} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.398396] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 738.398796] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.593973] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3de1c9ad-6b68-4184-a1e8-47d8df8163d3 tempest-ServersListShow296Test-275716723 tempest-ServersListShow296Test-275716723-project-member] Lock "8b478626-e3dc-42c4-bd59-fa0a71d798c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.979s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.736687] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52903ed8-3bb3-7e2b-0d06-e48369e8e40d, 'name': SearchDatastore_Task, 'duration_secs': 0.01351} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.739737] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 738.740043] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.740346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.740544] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.740775] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.741343] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.741690] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 738.741958] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74727e29-3ff0-473f-9301-9c60a3acb21f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.744108] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f34f3c1e-15a8-402d-9dab-435366b6849b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.750961] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 738.750961] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527933a4-60bd-a2ab-213b-6d2a1ec5bf16" [ 738.750961] env[65385]: _type = "Task" [ 738.750961] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.755362] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.755552] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 738.756797] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bef80efb-81ef-4f1d-82e7-c773b79e78b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.764977] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527933a4-60bd-a2ab-213b-6d2a1ec5bf16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.769508] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 738.769508] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a56372-29df-c097-9c41-969dc97616a5" [ 738.769508] env[65385]: _type = "Task" [ 738.769508] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.781024] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a56372-29df-c097-9c41-969dc97616a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.781149] env[65385]: DEBUG nova.compute.manager [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 738.782020] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adfb9ec-2df5-4f15-938b-2e28699e4803 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.184924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Acquiring lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.184924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Acquired lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.185223] env[65385]: DEBUG nova.network.neutron [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 739.195313] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f34860-d71d-4bb2-93fa-34e15346220d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.205283] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292c8908-36d2-4572-b814-a338bf3f99dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.241380] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e68a13e-e2c6-4ff8-9416-e438f505561f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.256397] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29e1e29-0644-4637-ac83-e7ff52800e50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.274796] env[65385]: DEBUG nova.compute.provider_tree [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.282778] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527933a4-60bd-a2ab-213b-6d2a1ec5bf16, 'name': SearchDatastore_Task, 'duration_secs': 0.040103} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.283456] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.283721] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 739.283949] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.292102] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a56372-29df-c097-9c41-969dc97616a5, 'name': SearchDatastore_Task, 'duration_secs': 0.031459} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.292917] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2e51fae-3ebe-4e5c-b81f-1acb362c233c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.297063] env[65385]: INFO nova.compute.manager [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] instance snapshotting [ 739.300690] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de70446-e7b8-4d47-894c-3c01de491c60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.305436] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 739.305436] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ee8731-52ef-f914-f3da-03bba140341b" [ 739.305436] env[65385]: _type = "Task" [ 739.305436] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.325158] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca53b62-4df3-4e82-827d-cf24cc2ac696 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.331603] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ee8731-52ef-f914-f3da-03bba140341b, 'name': SearchDatastore_Task, 'duration_secs': 0.01392} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.332577] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.332684] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 739.332912] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.333104] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.333339] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b99ef3d7-53cf-4604-b328-fbfeb98bfd61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.339475] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4ce0c72-2e0b-412e-ae9d-2afb618eb328 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.352375] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 739.352375] env[65385]: value = "task-4453477" [ 739.352375] env[65385]: _type = "Task" [ 739.352375] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.353895] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.354086] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 739.363110] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b70faea3-b0e2-4abf-bd5d-1160d69e4fea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.373105] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.374356] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 739.374356] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d5461-5b8c-7132-5741-0623a95a499b" [ 739.374356] env[65385]: _type = "Task" [ 739.374356] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.385396] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d5461-5b8c-7132-5741-0623a95a499b, 'name': SearchDatastore_Task, 'duration_secs': 0.010937} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.386405] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb2c3870-0a76-42da-8ac9-fe1b9382c508 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.393710] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 739.393710] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d241ff-91c8-4fde-f3e7-756b8958461b" [ 739.393710] env[65385]: _type = "Task" [ 739.393710] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.403822] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d241ff-91c8-4fde-f3e7-756b8958461b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.450279] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "26a2d464-0aab-463f-8265-948a4f0ff188" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 739.450571] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.653092] env[65385]: DEBUG nova.compute.manager [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Received event network-changed-ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 739.654879] env[65385]: DEBUG nova.compute.manager [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Refreshing instance network info cache due to event network-changed-ff2751e9-ab94-4f90-84f6-0d47f1514411. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 739.654879] env[65385]: DEBUG oslo_concurrency.lockutils [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Acquiring lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.654879] env[65385]: DEBUG oslo_concurrency.lockutils [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Acquired lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.654879] env[65385]: DEBUG nova.network.neutron [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Refreshing network info cache for port ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 739.687592] env[65385]: WARNING neutronclient.v2_0.client [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.688418] env[65385]: WARNING openstack [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.689288] env[65385]: WARNING openstack [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.784922] env[65385]: DEBUG nova.scheduler.client.report [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.825135] env[65385]: WARNING openstack [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.825456] env[65385]: WARNING openstack [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.848994] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 739.849608] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-eff4d577-cdf2-42a6-85e5-e1b483f6e555 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.859661] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 739.859661] env[65385]: value = "task-4453478" [ 739.859661] env[65385]: _type = "Task" [ 739.859661] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.869504] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453477, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.880995] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453478, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.904635] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d241ff-91c8-4fde-f3e7-756b8958461b, 'name': SearchDatastore_Task, 'duration_secs': 0.010936} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.906315] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.907250] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 739.907591] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d20e68f4-df8f-436a-a76c-7bef9c48fc98 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.912141] env[65385]: DEBUG nova.compute.manager [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Received event network-changed-ad95d76f-e023-47b2-a936-e7ed11779276 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 739.912141] env[65385]: DEBUG nova.compute.manager [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Refreshing instance network info cache due to event network-changed-ad95d76f-e023-47b2-a936-e7ed11779276. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 739.912141] env[65385]: DEBUG oslo_concurrency.lockutils [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Acquiring lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.912302] env[65385]: DEBUG oslo_concurrency.lockutils [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Acquired lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.913017] env[65385]: DEBUG nova.network.neutron [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Refreshing network info cache for port ad95d76f-e023-47b2-a936-e7ed11779276 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 739.923551] env[65385]: WARNING neutronclient.v2_0.client [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.925842] env[65385]: WARNING openstack [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.925842] env[65385]: WARNING openstack [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.938119] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 739.938119] env[65385]: value = "task-4453479" [ 739.938119] env[65385]: _type = "Task" [ 739.938119] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.955457] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453479, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.068676] env[65385]: DEBUG nova.network.neutron [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Updating instance_info_cache with network_info: [{"id": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "address": "fa:16:3e:91:29:10", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap884a69db-a8", "ovs_interfaceid": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 740.157351] env[65385]: WARNING neutronclient.v2_0.client [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.158453] env[65385]: WARNING openstack [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.159216] env[65385]: WARNING openstack [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.236617] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "ad29afd9-3e52-4fac-977f-91774acad7d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 740.237088] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.291841] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.241s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.297361] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.546s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.299825] env[65385]: INFO nova.compute.claims [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.363689] env[65385]: INFO nova.scheduler.client.report [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleted allocations for instance 1bc94ee7-d4f9-48c3-97f1-9e662eb50582 [ 740.376076] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453477, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52257} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.379733] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 740.380087] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 740.380380] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453478, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.380638] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c7a190d-c205-46d1-8307-5c9695b5306f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.387608] env[65385]: WARNING openstack [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.388044] env[65385]: WARNING openstack [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.398679] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 740.398679] env[65385]: value = "task-4453480" [ 740.398679] env[65385]: _type = "Task" [ 740.398679] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.412490] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453480, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.418475] env[65385]: WARNING neutronclient.v2_0.client [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.419029] env[65385]: WARNING openstack [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.419424] env[65385]: WARNING openstack [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.454978] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453479, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.531135] env[65385]: WARNING neutronclient.v2_0.client [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.531791] env[65385]: WARNING openstack [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.532158] env[65385]: WARNING openstack [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.560319] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "f25cc41e-560e-4cab-8bf0-dac47899d037" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 740.560587] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "f25cc41e-560e-4cab-8bf0-dac47899d037" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.572289] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Releasing lock "refresh_cache-34730159-3bec-4ad5-b85e-0f67998d6001" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.572698] env[65385]: DEBUG nova.compute.manager [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Inject network info {{(pid=65385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7777}} [ 740.572908] env[65385]: DEBUG nova.compute.manager [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] network_info to inject: |[{"id": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "address": "fa:16:3e:91:29:10", "network": {"id": "f8666347-a362-46c8-a867-614d988de655", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1541011363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9a1bf7c2dad4808a3782a3fba8a25fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6815237d-f565-474d-a3c0-9c675478eb00", "external-id": "nsx-vlan-transportzone-526", "segmentation_id": 526, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap884a69db-a8", "ovs_interfaceid": "884a69db-a84a-46ae-9af6-bd5f21a386e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7778}} [ 740.581292] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Reconfiguring VM instance to set the machine id {{(pid=65385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 740.589655] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4de6e77-4fc5-426b-b40d-309787875eef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.615263] env[65385]: DEBUG oslo_vmware.api [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Waiting for the task: (returnval){ [ 740.615263] env[65385]: value = "task-4453481" [ 740.615263] env[65385]: _type = "Task" [ 740.615263] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.638150] env[65385]: DEBUG oslo_vmware.api [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Task: {'id': task-4453481, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.639785] env[65385]: WARNING openstack [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.639785] env[65385]: WARNING openstack [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.678544] env[65385]: DEBUG nova.network.neutron [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updated VIF entry in instance network info cache for port ff2751e9-ab94-4f90-84f6-0d47f1514411. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 740.678911] env[65385]: DEBUG nova.network.neutron [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updating instance_info_cache with network_info: [{"id": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "address": "fa:16:3e:a0:d3:7d", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff2751e9-ab", "ovs_interfaceid": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 740.747115] env[65385]: WARNING neutronclient.v2_0.client [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.747939] env[65385]: WARNING openstack [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.747939] env[65385]: WARNING openstack [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.855627] env[65385]: DEBUG nova.network.neutron [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updated VIF entry in instance network info cache for port ad95d76f-e023-47b2-a936-e7ed11779276. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 740.856052] env[65385]: DEBUG nova.network.neutron [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance_info_cache with network_info: [{"id": "ad95d76f-e023-47b2-a936-e7ed11779276", "address": "fa:16:3e:b0:c7:d2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95d76f-e0", "ovs_interfaceid": "ad95d76f-e023-47b2-a936-e7ed11779276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 740.878603] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453478, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.879113] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8c1a6c9-61a7-426a-996a-84ce41260798 tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "1bc94ee7-d4f9-48c3-97f1-9e662eb50582" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.429s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.909962] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453480, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086855} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.910502] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.911289] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485cdb17-630e-4684-94a2-0bdbf9c43c56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.941562] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.943979] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f52ee313-ed65-4e48-a5e1-2a1a93d90c55 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.974892] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453479, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.977417] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 740.977417] env[65385]: value = "task-4453482" [ 740.977417] env[65385]: _type = "Task" [ 740.977417] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.988246] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453482, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.126214] env[65385]: DEBUG oslo_vmware.api [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] Task: {'id': task-4453481, 'name': ReconfigVM_Task, 'duration_secs': 0.316604} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.126641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9ba41d7a-86f6-498c-b91e-7b83c7123f37 tempest-ServersAdminTestJSON-138627310 tempest-ServersAdminTestJSON-138627310-project-admin] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Reconfigured VM instance to set the machine id {{(pid=65385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 741.129806] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "c798006c-3e74-461f-8397-3d8ca0be0a95" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.130166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.130555] env[65385]: INFO nova.compute.manager [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Rebooting instance [ 741.182571] env[65385]: DEBUG oslo_concurrency.lockutils [req-69be04f9-f52f-45e4-ad0c-25c22974580f req-59ff7325-7001-4cb2-a8d1-3e7d838036bb service nova] Releasing lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.359315] env[65385]: DEBUG oslo_concurrency.lockutils [req-26a3f9a6-0aef-4d26-b11e-f3f5877c6b6f req-c780ffc1-2fc4-485e-9a45-2d9a0de16b91 service nova] Releasing lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.385973] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453478, 'name': CreateSnapshot_Task, 'duration_secs': 1.172308} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.389114] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 741.390156] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e8735a-3a6d-4d99-be2c-2050d3845ea0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.466656] env[65385]: DEBUG nova.compute.manager [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 741.468043] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a52ca1-40a9-439f-8377-cacb598fbbcf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.480907] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453479, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.523225} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.485494] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 741.485796] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.488159] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed8b4581-e4db-434f-9181-4259acd5fdf1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.498636] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453482, 'name': ReconfigVM_Task, 'duration_secs': 0.38852} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.502852] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 741.504700] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 741.504700] env[65385]: value = "task-4453483" [ 741.504700] env[65385]: _type = "Task" [ 741.504700] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.505268] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d8666f4-b054-4ecf-b9e6-636dd293776e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.517905] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.522738] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 741.522738] env[65385]: value = "task-4453484" [ 741.522738] env[65385]: _type = "Task" [ 741.522738] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.537668] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453484, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.671077] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.671331] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquired lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.671556] env[65385]: DEBUG nova.network.neutron [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 741.915894] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 741.917034] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.917034] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.917034] env[65385]: DEBUG nova.compute.manager [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 741.917034] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-76f4385e-41c7-407b-a4a6-7bc1212f2484 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.921404] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dea2817-0149-479d-9daf-774f90167913 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.935633] env[65385]: DEBUG nova.compute.manager [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 741.935633] env[65385]: DEBUG nova.objects.instance [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'flavor' on Instance uuid 59f26752-75ab-45e5-b0a9-bbcef44ecd7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 741.939025] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 741.939025] env[65385]: value = "task-4453485" [ 741.939025] env[65385]: _type = "Task" [ 741.939025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.954016] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453485, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.996679] env[65385]: INFO nova.compute.manager [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] instance snapshotting [ 741.997707] env[65385]: DEBUG nova.objects.instance [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'flavor' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 742.018737] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076371} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.023070] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.024275] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4608626-33a5-464b-8abc-b8bcdbbf2498 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.058304] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.058671] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453484, 'name': Rename_Task, 'duration_secs': 0.1562} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.065832] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3846fa4f-55b6-465c-97a0-f165984fdb89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.082409] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 742.085333] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dca496a3-0ced-43f8-a4e8-9b2c503fd2fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.098625] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 742.098625] env[65385]: value = "task-4453486" [ 742.098625] env[65385]: _type = "Task" [ 742.098625] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.100400] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 742.100400] env[65385]: value = "task-4453487" [ 742.100400] env[65385]: _type = "Task" [ 742.100400] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.110708] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe10f2b0-c67e-4b14-972e-bbe79275592c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.122363] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453486, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.129308] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453487, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.131108] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243949d4-0543-496c-a218-eef31114810b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.172095] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973b752e-3c12-4d15-a441-7a5f46ba587e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.175820] env[65385]: WARNING neutronclient.v2_0.client [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.176623] env[65385]: WARNING openstack [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.177031] env[65385]: WARNING openstack [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.191790] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09dee6c8-4d98-4fa7-ba20-3e7bae0cfcdc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.209367] env[65385]: DEBUG nova.compute.provider_tree [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.460560] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453485, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.497845] env[65385]: WARNING openstack [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.498306] env[65385]: WARNING openstack [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.511058] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3cf41ee-d773-4215-9367-28b0934cb811 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.536226] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c56db4-7894-41c6-a5f1-7a8ea81e466a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.591886] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.592330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.592657] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.592917] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.593147] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.599955] env[65385]: INFO nova.compute.manager [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Terminating instance [ 742.629975] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453486, 'name': ReconfigVM_Task, 'duration_secs': 0.324579} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.637571] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 742.639095] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453487, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.642712] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4782d6a3-389d-4424-b8a1-25d72d9dac0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.652476] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 742.652476] env[65385]: value = "task-4453488" [ 742.652476] env[65385]: _type = "Task" [ 742.652476] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.667820] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453488, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.672098] env[65385]: WARNING neutronclient.v2_0.client [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.672098] env[65385]: WARNING openstack [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.672098] env[65385]: WARNING openstack [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.717307] env[65385]: DEBUG nova.scheduler.client.report [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 742.826927] env[65385]: DEBUG nova.network.neutron [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updating instance_info_cache with network_info: [{"id": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "address": "fa:16:3e:a0:d3:7d", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff2751e9-ab", "ovs_interfaceid": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.956243] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 742.956613] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453485, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.956895] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53df1907-91a2-4167-9793-a84d95735236 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.968141] env[65385]: DEBUG oslo_vmware.api [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 742.968141] env[65385]: value = "task-4453489" [ 742.968141] env[65385]: _type = "Task" [ 742.968141] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.984666] env[65385]: DEBUG oslo_vmware.api [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.049634] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 743.049885] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9d51f638-dde1-43bf-9e04-6144e51ebf53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.059815] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 743.059815] env[65385]: value = "task-4453490" [ 743.059815] env[65385]: _type = "Task" [ 743.059815] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.069849] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453490, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.108944] env[65385]: DEBUG nova.compute.manager [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 743.109992] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.110078] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a011f4-6c22-451d-a7ed-c5e8be286e6a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.122302] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 743.123103] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-518af248-b796-4d47-9cbe-ea48bdf995a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.128639] env[65385]: DEBUG oslo_vmware.api [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453487, 'name': PowerOnVM_Task, 'duration_secs': 0.541014} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.130077] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 743.130467] env[65385]: INFO nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Took 9.17 seconds to spawn the instance on the hypervisor. [ 743.130603] env[65385]: DEBUG nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 743.130972] env[65385]: DEBUG oslo_vmware.api [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 743.130972] env[65385]: value = "task-4453491" [ 743.130972] env[65385]: _type = "Task" [ 743.130972] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.132151] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae23b12-a073-4869-aa63-aa59bcf4402d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.146301] env[65385]: DEBUG oslo_vmware.api [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.163704] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453488, 'name': Rename_Task, 'duration_secs': 0.160169} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.165139] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 743.165139] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1b615ec-322a-4f39-bfa6-7a86b347c29c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.174946] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 743.174946] env[65385]: value = "task-4453492" [ 743.174946] env[65385]: _type = "Task" [ 743.174946] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.188409] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453492, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.227194] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.930s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.230104] env[65385]: DEBUG nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 743.232212] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.425s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.235439] env[65385]: INFO nova.compute.claims [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.330135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Releasing lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.455060] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453485, 'name': CloneVM_Task, 'duration_secs': 1.519933} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.455060] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Created linked-clone VM from snapshot [ 743.455928] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93b26cd-8456-4271-9c3a-c38d1e70b12a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.466140] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Uploading image 83526769-8e8f-421d-9b70-d3b7153a144b {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 743.479777] env[65385]: DEBUG oslo_vmware.api [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453489, 'name': PowerOffVM_Task, 'duration_secs': 0.233987} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.480098] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 743.480379] env[65385]: DEBUG nova.compute.manager [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 743.481364] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98ce30f-a5f8-40ce-a809-f9b9a0ac716d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.501408] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 743.501408] env[65385]: value = "vm-871005" [ 743.501408] env[65385]: _type = "VirtualMachine" [ 743.501408] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 743.502648] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ba9fb25c-f8e0-48a0-b288-7651ae73e42a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.510573] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lease: (returnval){ [ 743.510573] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fde581-5e8e-4c15-c4bc-c5862c3e3619" [ 743.510573] env[65385]: _type = "HttpNfcLease" [ 743.510573] env[65385]: } obtained for exporting VM: (result){ [ 743.510573] env[65385]: value = "vm-871005" [ 743.510573] env[65385]: _type = "VirtualMachine" [ 743.510573] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 743.510888] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the lease: (returnval){ [ 743.510888] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fde581-5e8e-4c15-c4bc-c5862c3e3619" [ 743.510888] env[65385]: _type = "HttpNfcLease" [ 743.510888] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 743.520856] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 743.520856] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fde581-5e8e-4c15-c4bc-c5862c3e3619" [ 743.520856] env[65385]: _type = "HttpNfcLease" [ 743.520856] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 743.571997] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453490, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.646168] env[65385]: DEBUG oslo_vmware.api [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453491, 'name': PowerOffVM_Task, 'duration_secs': 0.232461} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.646495] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 743.646724] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 743.647095] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2ddcf44-4a7f-48d9-a360-f80617f4c269 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.660363] env[65385]: INFO nova.compute.manager [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Took 40.18 seconds to build instance. [ 743.688277] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453492, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.721293] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 743.721919] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 743.721919] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleting the datastore file [datastore2] 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 743.722134] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-394c7cfe-e89c-44fe-8b29-8a7aa16e54f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.732022] env[65385]: DEBUG oslo_vmware.api [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 743.732022] env[65385]: value = "task-4453495" [ 743.732022] env[65385]: _type = "Task" [ 743.732022] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.743084] env[65385]: DEBUG nova.compute.utils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 743.747271] env[65385]: DEBUG oslo_vmware.api [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.747271] env[65385]: DEBUG nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 743.747271] env[65385]: DEBUG nova.network.neutron [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 743.747435] env[65385]: WARNING neutronclient.v2_0.client [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.747719] env[65385]: WARNING neutronclient.v2_0.client [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.748310] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.748688] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.840738] env[65385]: DEBUG nova.compute.manager [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 743.840738] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57291bf7-3e4b-4d63-a518-9ab0f308f74b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.855254] env[65385]: DEBUG nova.policy [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3979c16abd3b451db6dffaf7c2bad861', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1a38e5ebbe040a4840925352ca4c779', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 743.997530] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f4197682-7624-40aa-a542-be1b4988671a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.081s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.028321] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 744.028321] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fde581-5e8e-4c15-c4bc-c5862c3e3619" [ 744.028321] env[65385]: _type = "HttpNfcLease" [ 744.028321] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 744.028715] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 744.028715] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fde581-5e8e-4c15-c4bc-c5862c3e3619" [ 744.028715] env[65385]: _type = "HttpNfcLease" [ 744.028715] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 744.029784] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dceb772d-55f9-48ad-af00-77075a3438f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.045045] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a268c8-bbfb-f2d6-eed2-5b63d10d1dbf/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 744.045045] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a268c8-bbfb-f2d6-eed2-5b63d10d1dbf/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 744.129912] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453490, 'name': CreateSnapshot_Task, 'duration_secs': 1.01688} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.130204] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 744.131135] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d238c9b-7436-4214-9b9c-cff6a8eb9ecd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.164806] env[65385]: DEBUG oslo_concurrency.lockutils [None req-95e81247-f357-4d5b-8cbe-6e989e933c6a tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.694s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.192399] env[65385]: DEBUG oslo_vmware.api [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453492, 'name': PowerOnVM_Task, 'duration_secs': 0.692233} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.192979] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 744.193884] env[65385]: INFO nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Took 6.44 seconds to spawn the instance on the hypervisor. [ 744.193884] env[65385]: DEBUG nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 744.193884] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6f908dc0-811b-481b-aada-8f0f60f607aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.196614] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953152ca-9779-4ef3-880a-2c11ad20cabf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.221171] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.221590] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.248028] env[65385]: DEBUG nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 744.255207] env[65385]: DEBUG oslo_vmware.api [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453495, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288962} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.255937] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 744.256192] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 744.257073] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.257073] env[65385]: INFO nova.compute.manager [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Took 1.15 seconds to destroy the instance on the hypervisor. [ 744.257073] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 744.257311] env[65385]: DEBUG nova.compute.manager [-] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 744.257421] env[65385]: DEBUG nova.network.neutron [-] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 744.257652] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.258196] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.258442] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.387217] env[65385]: DEBUG nova.network.neutron [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Successfully created port: 0ca9c1b7-9149-4367-af50-7af1b8ef0051 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 744.435088] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.585907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266848b0-59a9-4a1b-9d73-7cbfe475987a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.598817] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a521a0-6066-447d-a7f4-174e62547b07 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.635797] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31f7d29-6f76-475b-9a4c-bf2424560832 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.645982] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127cc226-69d2-444a-bc4c-ae795f82b254 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.660288] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 744.660661] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-72629703-d112-40dd-8c12-78cb7d26927b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.676050] env[65385]: DEBUG nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 744.679193] env[65385]: DEBUG nova.compute.provider_tree [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.682756] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 744.682756] env[65385]: value = "task-4453496" [ 744.682756] env[65385]: _type = "Task" [ 744.682756] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.694187] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453496, 'name': CloneVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.729144] env[65385]: INFO nova.compute.manager [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Took 37.53 seconds to build instance. [ 744.867073] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a92db9-2563-4086-8124-80dc7defec1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.878869] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Doing hard reboot of VM {{(pid=65385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 744.879939] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-bb81c5ae-2131-459b-8342-af0fe847887f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.889199] env[65385]: DEBUG oslo_vmware.api [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 744.889199] env[65385]: value = "task-4453497" [ 744.889199] env[65385]: _type = "Task" [ 744.889199] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.904726] env[65385]: DEBUG oslo_vmware.api [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453497, 'name': ResetVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.185405] env[65385]: DEBUG nova.scheduler.client.report [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 745.205296] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453496, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.212778] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.231882] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d28d9bd2-ae8e-4449-9bd2-48efff366677 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.057s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 745.263816] env[65385]: DEBUG nova.network.neutron [-] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 745.267394] env[65385]: DEBUG nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 745.278984] env[65385]: DEBUG nova.compute.manager [req-302e2cca-a969-462f-be70-fc250e600e8e req-1a91354d-208f-4540-9b08-7c9d3382447e service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Received event network-vif-deleted-99c91233-f17e-45f0-8852-fd19e18a8502 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 745.278984] env[65385]: INFO nova.compute.manager [req-302e2cca-a969-462f-be70-fc250e600e8e req-1a91354d-208f-4540-9b08-7c9d3382447e service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Neutron deleted interface 99c91233-f17e-45f0-8852-fd19e18a8502; detaching it from the instance and deleting it from the info cache [ 745.278984] env[65385]: DEBUG nova.network.neutron [req-302e2cca-a969-462f-be70-fc250e600e8e req-1a91354d-208f-4540-9b08-7c9d3382447e service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.132049] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 746.132049] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 746.132049] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 746.132049] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 746.132049] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 746.132049] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 746.132049] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.132680] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 746.132718] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 746.133112] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 746.133404] env[65385]: DEBUG nova.virt.hardware [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 746.134589] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.903s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.135813] env[65385]: DEBUG nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 746.138890] env[65385]: DEBUG nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 746.141958] env[65385]: INFO nova.compute.manager [-] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Took 1.88 seconds to deallocate network for instance. [ 746.147318] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8ef872-43fd-4c20-b8ed-6e3e3b5c916b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.158249] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.895s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.158249] env[65385]: DEBUG nova.objects.instance [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lazy-loading 'resources' on Instance uuid 427d5c6d-ab01-42ef-8ee5-edd608896b8b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 746.160950] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4762141b-546b-4444-958d-72001066265b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.187471] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec40494-6084-4f21-aeea-10532a3c6949 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.193960] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453496, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.194528] env[65385]: DEBUG oslo_vmware.api [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453497, 'name': ResetVM_Task, 'duration_secs': 0.111797} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.196090] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Did hard reboot of VM {{(pid=65385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 746.196290] env[65385]: DEBUG nova.compute.manager [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 746.198839] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11b54f4-cdaa-4e83-be92-57a88bc71ed9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.208813] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730d6b7d-3e07-4438-a572-4198f76f6de6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.231495] env[65385]: DEBUG nova.network.neutron [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Successfully updated port: 0ca9c1b7-9149-4367-af50-7af1b8ef0051 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 746.281160] env[65385]: DEBUG nova.compute.manager [req-302e2cca-a969-462f-be70-fc250e600e8e req-1a91354d-208f-4540-9b08-7c9d3382447e service nova] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Detach interface failed, port_id=99c91233-f17e-45f0-8852-fd19e18a8502, reason: Instance 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 746.547129] env[65385]: DEBUG nova.compute.manager [req-759de0c0-c299-4a66-92e0-fb475f43e5c0 req-6849c004-a69d-497b-90da-1f29cc45a04f service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Received event network-vif-plugged-0ca9c1b7-9149-4367-af50-7af1b8ef0051 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 746.547389] env[65385]: DEBUG oslo_concurrency.lockutils [req-759de0c0-c299-4a66-92e0-fb475f43e5c0 req-6849c004-a69d-497b-90da-1f29cc45a04f service nova] Acquiring lock "fae095a8-a768-4263-b456-ed0745398755-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.547638] env[65385]: DEBUG oslo_concurrency.lockutils [req-759de0c0-c299-4a66-92e0-fb475f43e5c0 req-6849c004-a69d-497b-90da-1f29cc45a04f service nova] Lock "fae095a8-a768-4263-b456-ed0745398755-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.547938] env[65385]: DEBUG oslo_concurrency.lockutils [req-759de0c0-c299-4a66-92e0-fb475f43e5c0 req-6849c004-a69d-497b-90da-1f29cc45a04f service nova] Lock "fae095a8-a768-4263-b456-ed0745398755-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.548197] env[65385]: DEBUG nova.compute.manager [req-759de0c0-c299-4a66-92e0-fb475f43e5c0 req-6849c004-a69d-497b-90da-1f29cc45a04f service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] No waiting events found dispatching network-vif-plugged-0ca9c1b7-9149-4367-af50-7af1b8ef0051 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 746.548406] env[65385]: WARNING nova.compute.manager [req-759de0c0-c299-4a66-92e0-fb475f43e5c0 req-6849c004-a69d-497b-90da-1f29cc45a04f service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Received unexpected event network-vif-plugged-0ca9c1b7-9149-4367-af50-7af1b8ef0051 for instance with vm_state building and task_state spawning. [ 746.649203] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453496, 'name': CloneVM_Task, 'duration_secs': 1.645882} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.649654] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Created linked-clone VM from snapshot [ 746.651237] env[65385]: DEBUG nova.compute.utils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 746.655605] env[65385]: DEBUG nova.objects.instance [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'flavor' on Instance uuid 59f26752-75ab-45e5-b0a9-bbcef44ecd7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 746.659275] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63414ef2-92bd-4a0b-92e2-d2f6ce1c13ae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.661944] env[65385]: DEBUG nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 746.663350] env[65385]: DEBUG nova.network.neutron [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 746.663793] env[65385]: WARNING neutronclient.v2_0.client [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.664171] env[65385]: WARNING neutronclient.v2_0.client [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.664933] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.665635] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.686164] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.698235] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Uploading image 7b43aa8d-6c77-4694-9308-fdf00d082211 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 746.707697] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.739241] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.739529] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquired lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 746.739679] env[65385]: DEBUG nova.network.neutron [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 746.746931] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 746.746931] env[65385]: value = "vm-871007" [ 746.746931] env[65385]: _type = "VirtualMachine" [ 746.746931] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 746.746931] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5431dc48-f7d8-4d45-b375-b11c740da04d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.748835] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eebfdaaa-17df-44fb-9a73-518435c3f7a8 tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.618s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.753584] env[65385]: DEBUG nova.policy [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27a258a335284ab3b73a6c55dfde1d6c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2de8098800694299aae5aa2f59f405bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 746.766614] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease: (returnval){ [ 746.766614] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525cb681-5d05-9b9e-576e-30e6340c5c10" [ 746.766614] env[65385]: _type = "HttpNfcLease" [ 746.766614] env[65385]: } obtained for exporting VM: (result){ [ 746.766614] env[65385]: value = "vm-871007" [ 746.766614] env[65385]: _type = "VirtualMachine" [ 746.766614] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 746.767088] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the lease: (returnval){ [ 746.767088] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525cb681-5d05-9b9e-576e-30e6340c5c10" [ 746.767088] env[65385]: _type = "HttpNfcLease" [ 746.767088] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 746.784554] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 746.784554] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525cb681-5d05-9b9e-576e-30e6340c5c10" [ 746.784554] env[65385]: _type = "HttpNfcLease" [ 746.784554] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 747.127543] env[65385]: DEBUG nova.compute.manager [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 747.166258] env[65385]: DEBUG nova.compute.utils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 747.172091] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.172180] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.173675] env[65385]: DEBUG nova.network.neutron [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 747.173675] env[65385]: DEBUG nova.objects.instance [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'info_cache' on Instance uuid 59f26752-75ab-45e5-b0a9-bbcef44ecd7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 747.239962] env[65385]: DEBUG nova.network.neutron [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Successfully created port: 0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 747.245742] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.246567] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.277654] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 747.277654] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525cb681-5d05-9b9e-576e-30e6340c5c10" [ 747.277654] env[65385]: _type = "HttpNfcLease" [ 747.277654] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 747.278232] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 747.278232] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525cb681-5d05-9b9e-576e-30e6340c5c10" [ 747.278232] env[65385]: _type = "HttpNfcLease" [ 747.278232] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 747.279211] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7f2901-92a4-4191-a715-380f9e5e233b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.292644] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ed86d2-e24a-2844-d596-9a2c6c5ea1c0/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 747.293240] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ed86d2-e24a-2844-d596-9a2c6c5ea1c0/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 747.382294] env[65385]: DEBUG nova.network.neutron [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 747.445495] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-abb25597-ac78-4e63-b7d4-4e070f63c6ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.486596] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2555d429-8560-408e-975d-b03c11e70301 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.490808] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.491231] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.503886] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130eecc3-a044-4eb4-b669-06859cecd9e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.539188] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954a4b2a-87f7-44b9-b5f3-ea1b8250e2c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.548906] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5dfb32-f293-410c-88dc-18bbdd321a0d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.564860] env[65385]: DEBUG nova.compute.provider_tree [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.660498] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.676033] env[65385]: DEBUG nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 747.680082] env[65385]: DEBUG nova.objects.base [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Object Instance<59f26752-75ab-45e5-b0a9-bbcef44ecd7f> lazy-loaded attributes: flavor,info_cache {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 747.755861] env[65385]: WARNING neutronclient.v2_0.client [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 747.756874] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.757327] env[65385]: WARNING openstack [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.978467] env[65385]: DEBUG nova.network.neutron [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Updating instance_info_cache with network_info: [{"id": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "address": "fa:16:3e:9a:8f:26", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ca9c1b7-91", "ovs_interfaceid": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 748.068531] env[65385]: DEBUG nova.scheduler.client.report [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 748.187050] env[65385]: WARNING neutronclient.v2_0.client [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.188984] env[65385]: WARNING openstack [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.188984] env[65385]: WARNING openstack [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.379902] env[65385]: WARNING openstack [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.380532] env[65385]: WARNING openstack [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.481057] env[65385]: WARNING neutronclient.v2_0.client [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.481821] env[65385]: WARNING openstack [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.482258] env[65385]: WARNING openstack [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.492995] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Releasing lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.493842] env[65385]: DEBUG nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Instance network_info: |[{"id": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "address": "fa:16:3e:9a:8f:26", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ca9c1b7-91", "ovs_interfaceid": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 748.495340] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:8f:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0ca9c1b7-9149-4367-af50-7af1b8ef0051', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 748.503822] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Creating folder: Project (d1a38e5ebbe040a4840925352ca4c779). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 748.503995] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d50d868-57a2-4b08-9826-e73d587d83f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.517949] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Created folder: Project (d1a38e5ebbe040a4840925352ca4c779) in parent group-v870881. [ 748.518401] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Creating folder: Instances. Parent ref: group-v871008. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 748.518699] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-263e2332-06b3-442b-a17b-577f48196be8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.534141] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Created folder: Instances in parent group-v871008. [ 748.534696] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 748.535022] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fae095a8-a768-4263-b456-ed0745398755] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 748.535365] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07384629-73ff-4158-aa6d-092b05d65734 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.566098] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 748.566098] env[65385]: value = "task-4453501" [ 748.566098] env[65385]: _type = "Task" [ 748.566098] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.576559] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.417s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.582223] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.037s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.582767] env[65385]: DEBUG nova.objects.instance [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'resources' on Instance uuid 28fd3828-38ec-426f-b4dc-1e094304780b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.584658] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453501, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.603530] env[65385]: INFO nova.scheduler.client.report [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Deleted allocations for instance 427d5c6d-ab01-42ef-8ee5-edd608896b8b [ 748.664383] env[65385]: DEBUG nova.network.neutron [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 748.697555] env[65385]: DEBUG nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 749.078847] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453501, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.114969] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b384cd3e-323d-4b92-a119-7957f9c48e3b tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "427d5c6d-ab01-42ef-8ee5-edd608896b8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.316s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.169465] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 749.395876] env[65385]: DEBUG nova.network.neutron [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Successfully updated port: 0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 749.580194] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453501, 'name': CreateVM_Task, 'duration_secs': 0.519992} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.583765] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fae095a8-a768-4263-b456-ed0745398755] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 749.584479] env[65385]: WARNING neutronclient.v2_0.client [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.584852] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.584982] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.585398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 749.585786] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8720a5e0-cbd2-4bef-bbe7-b5c5b1a91112 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.593385] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 749.593385] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d8bb03-12cc-85a4-8f34-3b090c4827b8" [ 749.593385] env[65385]: _type = "Task" [ 749.593385] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.606822] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d8bb03-12cc-85a4-8f34-3b090c4827b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.734054] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d041242d-2ce9-4c84-8786-b7bd2527b349 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.743600] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc097da-f394-4d29-8622-a545080984f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.778572] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735945e9-1412-47da-8c39-687e27b8505e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.787547] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d35a0b-293a-4098-97a8-19a30cbfc940 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.802755] env[65385]: DEBUG nova.compute.provider_tree [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.899604] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.899791] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.899994] env[65385]: DEBUG nova.network.neutron [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 750.107895] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d8bb03-12cc-85a4-8f34-3b090c4827b8, 'name': SearchDatastore_Task, 'duration_secs': 0.016847} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.108284] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.108543] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.108779] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.108932] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.109113] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.109407] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcdd841e-11c1-46c3-9f6a-41993cbe128b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.120388] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 750.120574] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 750.121369] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72bcb17d-5c52-453a-9255-8ad7b6e482e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.128290] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 750.128290] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525adebc-68fd-4141-c51d-7503f1ef528b" [ 750.128290] env[65385]: _type = "Task" [ 750.128290] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.138612] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525adebc-68fd-4141-c51d-7503f1ef528b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.178126] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 750.178461] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b17e992-031a-4712-aa87-49a41121cd4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.186663] env[65385]: DEBUG oslo_vmware.api [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 750.186663] env[65385]: value = "task-4453502" [ 750.186663] env[65385]: _type = "Task" [ 750.186663] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.196539] env[65385]: DEBUG oslo_vmware.api [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453502, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.306170] env[65385]: DEBUG nova.scheduler.client.report [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 750.403671] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.404342] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.452015] env[65385]: DEBUG nova.network.neutron [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 750.475635] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.476136] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.554925] env[65385]: WARNING neutronclient.v2_0.client [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.555948] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.556496] env[65385]: WARNING openstack [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.640597] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525adebc-68fd-4141-c51d-7503f1ef528b, 'name': SearchDatastore_Task, 'duration_secs': 0.014885} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.641486] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-588d2536-828d-497a-b3ea-8b6bede1f82f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.648193] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 750.648193] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a28ca0-5a40-9d24-a65f-726a20f174a8" [ 750.648193] env[65385]: _type = "Task" [ 750.648193] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.659303] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a28ca0-5a40-9d24-a65f-726a20f174a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.662362] env[65385]: DEBUG nova.network.neutron [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Updating instance_info_cache with network_info: [{"id": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "address": "fa:16:3e:93:c1:c4", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ee1cb97-07", "ovs_interfaceid": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.698748] env[65385]: DEBUG oslo_vmware.api [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453502, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.811501] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.229s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.813979] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.563s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.815638] env[65385]: INFO nova.compute.claims [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.834801] env[65385]: INFO nova.scheduler.client.report [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted allocations for instance 28fd3828-38ec-426f-b4dc-1e094304780b [ 751.160733] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a28ca0-5a40-9d24-a65f-726a20f174a8, 'name': SearchDatastore_Task, 'duration_secs': 0.01701} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.161042] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.161391] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] fae095a8-a768-4263-b456-ed0745398755/fae095a8-a768-4263-b456-ed0745398755.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 751.161679] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57034efa-be84-45af-b377-4893585a6caa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.164776] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Releasing lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.165313] env[65385]: DEBUG nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Instance network_info: |[{"id": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "address": "fa:16:3e:93:c1:c4", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ee1cb97-07", "ovs_interfaceid": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 751.170968] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 751.170968] env[65385]: value = "task-4453503" [ 751.170968] env[65385]: _type = "Task" [ 751.170968] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.181143] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453503, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.200040] env[65385]: DEBUG oslo_vmware.api [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453502, 'name': PowerOnVM_Task, 'duration_secs': 0.590988} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.200040] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 751.201066] env[65385]: DEBUG nova.compute.manager [None req-4616e197-4c99-4e19-a297-742f8381c494 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 751.201668] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43015678-c88e-409d-aa3b-6c7fd9d95341 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.343312] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87516f49-807e-42a2-8473-19db1df0daec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "28fd3828-38ec-426f-b4dc-1e094304780b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.280s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.687517] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453503, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.186963] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453503, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721783} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.187322] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] fae095a8-a768-4263-b456-ed0745398755/fae095a8-a768-4263-b456-ed0745398755.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 752.187568] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 752.187842] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-346d7e73-f597-4409-9aa0-001f7c8d0550 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.199017] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 752.199017] env[65385]: value = "task-4453504" [ 752.199017] env[65385]: _type = "Task" [ 752.199017] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.209746] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453504, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.388357] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbc6b02-3ca2-4c60-bdd2-178b5a9b0469 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.397880] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d198510-53e3-4d85-84c1-69809ae938d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.436901] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb344c4c-d3f1-4dad-b762-7de3ea256333 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.446279] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f23e14-84c0-40ef-8b5f-1ae4632205f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.463180] env[65385]: DEBUG nova.compute.provider_tree [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.710104] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453504, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197093} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.710480] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.711473] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a72816-0548-4eff-8375-aebed1cc9c2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.735708] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] fae095a8-a768-4263-b456-ed0745398755/fae095a8-a768-4263-b456-ed0745398755.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.736516] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ccf5e2a-33a0-4183-b21d-86389de7aabb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.758061] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 752.758061] env[65385]: value = "task-4453505" [ 752.758061] env[65385]: _type = "Task" [ 752.758061] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.767834] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453505, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.967471] env[65385]: DEBUG nova.scheduler.client.report [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 753.268719] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453505, 'name': ReconfigVM_Task, 'duration_secs': 0.435966} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.268947] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Reconfigured VM instance instance-0000002b to attach disk [datastore2] fae095a8-a768-4263-b456-ed0745398755/fae095a8-a768-4263-b456-ed0745398755.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 753.269626] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32436987-4e11-40f3-b8fc-e6468c1cc188 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.279186] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 753.279186] env[65385]: value = "task-4453506" [ 753.279186] env[65385]: _type = "Task" [ 753.279186] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.289171] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453506, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.474260] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.475029] env[65385]: DEBUG nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 753.478420] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.774s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.478735] env[65385]: DEBUG nova.objects.instance [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lazy-loading 'resources' on Instance uuid 54ae4cd5-76d2-4e76-9528-6f511c4dca23 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 753.791160] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453506, 'name': Rename_Task, 'duration_secs': 0.296804} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.791582] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 753.791813] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e213087-06c2-4dba-b7d7-5ef64c1bd077 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.799076] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 753.799076] env[65385]: value = "task-4453507" [ 753.799076] env[65385]: _type = "Task" [ 753.799076] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.808262] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.982272] env[65385]: DEBUG nova.compute.utils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 753.987675] env[65385]: DEBUG nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 753.987901] env[65385]: DEBUG nova.network.neutron [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 753.988630] env[65385]: WARNING neutronclient.v2_0.client [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.988630] env[65385]: WARNING neutronclient.v2_0.client [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.989333] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.989715] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.069304] env[65385]: DEBUG nova.policy [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de6370f100c842a4875bc6639fbc1f27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f9a3a63deb9427bb5a08c7785509709', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 754.314272] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453507, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.416945] env[65385]: DEBUG nova.network.neutron [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Successfully created port: b74e7138-b571-449e-af1d-9e68bf8e9064 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 754.489236] env[65385]: DEBUG nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 754.555665] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27403e2-0c77-4bea-97f4-07576031867b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.565381] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80408abf-00ed-4646-9c64-49f973613bf5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.601262] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0a0ab1-0134-468f-832a-118960f3fd9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.611469] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b554f44-45f0-4194-86d0-5950e6ee9982 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.628303] env[65385]: DEBUG nova.compute.provider_tree [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.813086] env[65385]: DEBUG oslo_vmware.api [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453507, 'name': PowerOnVM_Task, 'duration_secs': 0.707706} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.813426] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 754.813650] env[65385]: INFO nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Took 9.55 seconds to spawn the instance on the hypervisor. [ 754.813860] env[65385]: DEBUG nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 754.814646] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fadb399-ed8b-4f2c-96ba-4a77b0b93f1a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.080708] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:47:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='1521481196',id=19,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-1727121178',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 755.080708] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 755.080708] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 755.080708] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 755.080900] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 755.080937] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 755.081213] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.081385] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 755.081550] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 755.081708] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 755.081875] env[65385]: DEBUG nova.virt.hardware [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 755.084489] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82267b1b-e116-4a82-949e-e33b60dfe5bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.094683] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07edb8ed-fabb-4f74-b732-df1ae3fb58a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.102102] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a268c8-bbfb-f2d6-eed2-5b63d10d1dbf/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 755.103362] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38eb9615-d58c-4e69-b5d4-b86d64f34000 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.110567] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a268c8-bbfb-f2d6-eed2-5b63d10d1dbf/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 755.110780] env[65385]: ERROR oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a268c8-bbfb-f2d6-eed2-5b63d10d1dbf/disk-0.vmdk due to incomplete transfer. [ 755.119451] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-122ae120-368b-4709-ad50-420bda7bfc9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.121684] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:c1:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f68ebd2a-3c68-48db-8c32-8a01497fc2e7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0ee1cb97-07c8-4c45-9442-d76deacb6028', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 755.129743] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 755.130085] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 755.130947] env[65385]: DEBUG nova.scheduler.client.report [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 755.134615] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bb7ab05-3b1e-43b3-99c8-9f04619c755a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.153096] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a268c8-bbfb-f2d6-eed2-5b63d10d1dbf/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 755.153344] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Uploaded image 83526769-8e8f-421d-9b70-d3b7153a144b to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 755.156257] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 755.157868] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8ff2581b-a37a-4da7-8890-bbf6d8422cd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.162973] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.162973] env[65385]: value = "task-4453508" [ 755.162973] env[65385]: _type = "Task" [ 755.162973] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.168284] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 755.168284] env[65385]: value = "task-4453509" [ 755.168284] env[65385]: _type = "Task" [ 755.168284] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.175558] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453508, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.185728] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453509, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.334622] env[65385]: INFO nova.compute.manager [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Took 43.61 seconds to build instance. [ 755.500927] env[65385]: DEBUG nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 755.532875] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 755.533182] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 755.533347] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 755.533690] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 755.533757] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 755.533871] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 755.534155] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.534291] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 755.534506] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 755.534673] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 755.534883] env[65385]: DEBUG nova.virt.hardware [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 755.535932] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeec4f57-fbd0-466b-98fe-674571574d2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.546469] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c99525c-5d3e-45d4-a9d5-ee25ba57be0c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.653011] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.174s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.657167] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.583s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.658616] env[65385]: INFO nova.compute.claims [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.682838] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453508, 'name': CreateVM_Task, 'duration_secs': 0.409585} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.684194] env[65385]: INFO nova.scheduler.client.report [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted allocations for instance 54ae4cd5-76d2-4e76-9528-6f511c4dca23 [ 755.686191] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 755.690306] env[65385]: INFO nova.compute.manager [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Rebuilding instance [ 755.692681] env[65385]: WARNING neutronclient.v2_0.client [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.693385] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.694025] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.694172] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 755.699065] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb34d477-1446-44f2-b6cb-094fe110875f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.702458] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453509, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.706565] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 755.706565] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ddace7-fd00-aff7-0fd7-e300f3097e26" [ 755.706565] env[65385]: _type = "Task" [ 755.706565] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.717816] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ddace7-fd00-aff7-0fd7-e300f3097e26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.746266] env[65385]: DEBUG nova.compute.manager [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 755.747811] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfb8120-b60c-46e6-8b04-70bd241fa883 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.837094] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eef0a652-08c9-429a-ae17-abc08e80f85f tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "fae095a8-a768-4263-b456-ed0745398755" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.122s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.936875] env[65385]: DEBUG nova.compute.manager [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Received event network-changed-0ca9c1b7-9149-4367-af50-7af1b8ef0051 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 755.937191] env[65385]: DEBUG nova.compute.manager [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Refreshing instance network info cache due to event network-changed-0ca9c1b7-9149-4367-af50-7af1b8ef0051. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 755.937507] env[65385]: DEBUG oslo_concurrency.lockutils [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Acquiring lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.937713] env[65385]: DEBUG oslo_concurrency.lockutils [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Acquired lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.937937] env[65385]: DEBUG nova.network.neutron [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Refreshing network info cache for port 0ca9c1b7-9149-4367-af50-7af1b8ef0051 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 756.035831] env[65385]: DEBUG nova.network.neutron [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Successfully updated port: b74e7138-b571-449e-af1d-9e68bf8e9064 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 756.181976] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453509, 'name': Destroy_Task, 'duration_secs': 0.71443} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.181976] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Destroyed the VM [ 756.181976] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 756.181976] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d60db56c-cabe-4593-aac9-0e724360a458 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.189552] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 756.189552] env[65385]: value = "task-4453510" [ 756.189552] env[65385]: _type = "Task" [ 756.189552] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.201509] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453510, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.201945] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fae9b83e-a1de-4d00-8567-7bf489a275a1 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "54ae4cd5-76d2-4e76-9528-6f511c4dca23" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.929s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.218361] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ddace7-fd00-aff7-0fd7-e300f3097e26, 'name': SearchDatastore_Task, 'duration_secs': 0.020295} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.219387] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.219627] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 756.219872] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.219999] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.220233] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.220758] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7c9dfee-9a4a-49c8-9fba-9bcdedd51fd0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.232091] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 756.232333] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 756.233105] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ef23a80-9d38-49b9-86d9-0ffecce16429 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.241924] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 756.241924] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528bdcd8-dd83-7748-6b7b-34b93f663d24" [ 756.241924] env[65385]: _type = "Task" [ 756.241924] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.251481] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528bdcd8-dd83-7748-6b7b-34b93f663d24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.339861] env[65385]: DEBUG nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 756.445226] env[65385]: WARNING neutronclient.v2_0.client [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.446050] env[65385]: WARNING openstack [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.446447] env[65385]: WARNING openstack [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.454433] env[65385]: DEBUG nova.compute.manager [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Received event network-changed-ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 756.454638] env[65385]: DEBUG nova.compute.manager [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Refreshing instance network info cache due to event network-changed-ff2751e9-ab94-4f90-84f6-0d47f1514411. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 756.454827] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Acquiring lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.454964] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Acquired lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.455165] env[65385]: DEBUG nova.network.neutron [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Refreshing network info cache for port ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 756.538635] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "refresh_cache-4232f3a4-c42b-4b9c-8e1d-da591634ffbb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.538870] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired lock "refresh_cache-4232f3a4-c42b-4b9c-8e1d-da591634ffbb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.539092] env[65385]: DEBUG nova.network.neutron [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 756.638203] env[65385]: WARNING openstack [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.638854] env[65385]: WARNING openstack [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.686641] env[65385]: DEBUG nova.compute.manager [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] [instance: fae095a8-a768-4263-b456-ed0745398755] Received event network-changed {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 756.687909] env[65385]: DEBUG nova.compute.manager [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] [instance: fae095a8-a768-4263-b456-ed0745398755] Refreshing instance network info cache due to event network-changed. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 756.687909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Acquiring lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.707808] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453510, 'name': RemoveSnapshot_Task} progress is 50%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.723594] env[65385]: WARNING neutronclient.v2_0.client [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.723861] env[65385]: WARNING openstack [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.724670] env[65385]: WARNING openstack [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.758922] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528bdcd8-dd83-7748-6b7b-34b93f663d24, 'name': SearchDatastore_Task, 'duration_secs': 0.02048} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.759891] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-211949b3-1537-41e2-9bfc-eb7f5269c375 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.763185] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 756.763450] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-113ffc93-c4a0-4c53-a015-fb4914a6287d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.773199] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 756.773199] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6ccdc-4912-6eb8-1c10-9e8cc67917f2" [ 756.773199] env[65385]: _type = "Task" [ 756.773199] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.779555] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 756.779555] env[65385]: value = "task-4453511" [ 756.779555] env[65385]: _type = "Task" [ 756.779555] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.791012] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6ccdc-4912-6eb8-1c10-9e8cc67917f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.797775] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.855258] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.855505] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.855687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.855945] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.857791] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.860489] env[65385]: DEBUG nova.network.neutron [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Updated VIF entry in instance network info cache for port 0ca9c1b7-9149-4367-af50-7af1b8ef0051. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 756.860939] env[65385]: DEBUG nova.network.neutron [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Updating instance_info_cache with network_info: [{"id": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "address": "fa:16:3e:9a:8f:26", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ca9c1b7-91", "ovs_interfaceid": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.866615] env[65385]: INFO nova.compute.manager [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Terminating instance [ 756.886395] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.959205] env[65385]: WARNING neutronclient.v2_0.client [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.959578] env[65385]: WARNING openstack [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.959976] env[65385]: WARNING openstack [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.046124] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.046711] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.118192] env[65385]: DEBUG nova.network.neutron [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 757.186963] env[65385]: WARNING openstack [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.187415] env[65385]: WARNING openstack [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.204606] env[65385]: DEBUG oslo_vmware.api [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453510, 'name': RemoveSnapshot_Task, 'duration_secs': 0.943584} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.204929] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 757.205123] env[65385]: INFO nova.compute.manager [None req-7371bfc1-ab75-46dc-a53e-bc24721543de tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Took 17.91 seconds to snapshot the instance on the hypervisor. [ 757.213213] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.213650] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.289390] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6ccdc-4912-6eb8-1c10-9e8cc67917f2, 'name': SearchDatastore_Task, 'duration_secs': 0.018348} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.290125] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.290428] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0/6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 757.290718] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d82a4735-4a8d-4573-8fce-fca77731dd97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.298021] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453511, 'name': PowerOffVM_Task, 'duration_secs': 0.327767} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.300835] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 757.301118] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.302669] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9c5d6e-1e50-4514-946f-396e933f04e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.308449] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 757.308449] env[65385]: value = "task-4453512" [ 757.308449] env[65385]: _type = "Task" [ 757.308449] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.317563] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 757.318044] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67426aaa-ba24-4f0c-9a1a-d28d8c3219c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.322026] env[65385]: WARNING neutronclient.v2_0.client [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.322026] env[65385]: WARNING openstack [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.322461] env[65385]: WARNING openstack [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.339688] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.351243] env[65385]: WARNING neutronclient.v2_0.client [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.352541] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.353255] env[65385]: WARNING openstack [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.367216] env[65385]: INFO nova.compute.manager [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Rebuilding instance [ 757.369448] env[65385]: DEBUG oslo_concurrency.lockutils [req-b8fc0567-5634-4d5b-b009-df7589178c84 req-aa4051d6-2946-4dfb-b937-b9d58fcb6c8e service nova] Releasing lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.370775] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Acquired lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.370984] env[65385]: DEBUG nova.network.neutron [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] [instance: fae095a8-a768-4263-b456-ed0745398755] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 757.375514] env[65385]: DEBUG nova.compute.manager [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 757.375804] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.377052] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a147093b-a51f-4280-bfb4-3ec820c88da2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.388715] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 757.390334] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-567770a1-5814-46df-aaea-7397d21d9b96 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.403250] env[65385]: DEBUG oslo_vmware.api [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 757.403250] env[65385]: value = "task-4453514" [ 757.403250] env[65385]: _type = "Task" [ 757.403250] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.423312] env[65385]: DEBUG oslo_vmware.api [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.450535] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 757.450777] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 757.452599] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleting the datastore file [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 757.452599] env[65385]: DEBUG nova.compute.manager [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 757.452599] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fc0cca4-1616-4ff8-a2c2-8feaadfd547a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.455672] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2753071f-ecfb-443b-8057-3920a4cdcf29 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.485780] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 757.485780] env[65385]: value = "task-4453515" [ 757.485780] env[65385]: _type = "Task" [ 757.485780] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.496923] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453515, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.499226] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce07ecb-3a39-405c-90e5-6b1736a23be8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.509719] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa742b8-1b44-46ed-9435-6528c44b3cee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.550547] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546d74d8-dd38-4a81-b2bf-fabbd837030c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.562668] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a54119-fd20-49fb-a672-7e47cbc322ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.573156] env[65385]: DEBUG nova.network.neutron [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Updating instance_info_cache with network_info: [{"id": "b74e7138-b571-449e-af1d-9e68bf8e9064", "address": "fa:16:3e:72:0c:9d", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.90", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb74e7138-b5", "ovs_interfaceid": "b74e7138-b571-449e-af1d-9e68bf8e9064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.585840] env[65385]: DEBUG nova.compute.provider_tree [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.633553] env[65385]: DEBUG nova.network.neutron [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updated VIF entry in instance network info cache for port ff2751e9-ab94-4f90-84f6-0d47f1514411. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 757.634124] env[65385]: DEBUG nova.network.neutron [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updating instance_info_cache with network_info: [{"id": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "address": "fa:16:3e:a0:d3:7d", "network": {"id": "9c9feda2-6497-4e6d-9390-2080aae50958", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-286905727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e94f2a868e2d4d31af54f2e5d8c479f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff2751e9-ab", "ovs_interfaceid": "ff2751e9-ab94-4f90-84f6-0d47f1514411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.681326] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ed86d2-e24a-2844-d596-9a2c6c5ea1c0/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 757.683105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82024806-fb80-46eb-9b42-18ded34c0327 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.691389] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ed86d2-e24a-2844-d596-9a2c6c5ea1c0/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 757.691591] env[65385]: ERROR oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ed86d2-e24a-2844-d596-9a2c6c5ea1c0/disk-0.vmdk due to incomplete transfer. [ 757.691921] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b8af5fb7-f2c1-4954-9360-d731f500f914 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.715592] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ed86d2-e24a-2844-d596-9a2c6c5ea1c0/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 757.715934] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Uploaded image 7b43aa8d-6c77-4694-9308-fdf00d082211 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 757.719041] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 757.719460] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-815e7f48-a57a-43ea-8d8e-9b240dcca22a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.731414] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "c798006c-3e74-461f-8397-3d8ca0be0a95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.732198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.732198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "c798006c-3e74-461f-8397-3d8ca0be0a95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.733051] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.733051] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.736431] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 757.736431] env[65385]: value = "task-4453516" [ 757.736431] env[65385]: _type = "Task" [ 757.736431] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.736431] env[65385]: INFO nova.compute.manager [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Terminating instance [ 757.752725] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453516, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.760533] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "fae095a8-a768-4263-b456-ed0745398755" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.760819] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "fae095a8-a768-4263-b456-ed0745398755" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.761268] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "fae095a8-a768-4263-b456-ed0745398755-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.761487] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "fae095a8-a768-4263-b456-ed0745398755-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.764568] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "fae095a8-a768-4263-b456-ed0745398755-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.764950] env[65385]: INFO nova.compute.manager [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Terminating instance [ 757.832425] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453512, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.876935] env[65385]: WARNING neutronclient.v2_0.client [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.877899] env[65385]: WARNING openstack [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.878462] env[65385]: WARNING openstack [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.915874] env[65385]: DEBUG oslo_vmware.api [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453514, 'name': PowerOffVM_Task, 'duration_secs': 0.246569} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.916097] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 757.916719] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 757.917222] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9be57c58-7749-4435-8b7f-77f96f1669c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.000890] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 758.001433] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 758.001675] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Deleting the datastore file [datastore1] 2e4c064f-27b0-4ccb-8fe6-c34a61153663 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 758.002048] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3adb26f0-4348-4436-ad23-29930e365de6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.013700] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453515, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.522138} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.015957] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 758.016264] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 758.016470] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.022737] env[65385]: DEBUG oslo_vmware.api [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for the task: (returnval){ [ 758.022737] env[65385]: value = "task-4453518" [ 758.022737] env[65385]: _type = "Task" [ 758.022737] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.036475] env[65385]: DEBUG oslo_vmware.api [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.089734] env[65385]: DEBUG nova.scheduler.client.report [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 758.094797] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Releasing lock "refresh_cache-4232f3a4-c42b-4b9c-8e1d-da591634ffbb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.095487] env[65385]: DEBUG nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Instance network_info: |[{"id": "b74e7138-b571-449e-af1d-9e68bf8e9064", "address": "fa:16:3e:72:0c:9d", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.90", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb74e7138-b5", "ovs_interfaceid": "b74e7138-b571-449e-af1d-9e68bf8e9064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 758.096956] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:0c:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b74e7138-b571-449e-af1d-9e68bf8e9064', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 758.107383] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 758.108058] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 758.108335] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-833c0bea-4f63-494a-98f1-5bd32d764b03 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.144348] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae64975d-b9dd-41f7-b319-2e0f3c8d2529 req-9973a0c9-323a-4585-9a18-d1f99ef14376 service nova] Releasing lock "refresh_cache-c798006c-3e74-461f-8397-3d8ca0be0a95" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.152071] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 758.152071] env[65385]: value = "task-4453519" [ 758.152071] env[65385]: _type = "Task" [ 758.152071] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.164945] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453519, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.169820] env[65385]: WARNING openstack [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.169820] env[65385]: WARNING openstack [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.249377] env[65385]: DEBUG nova.compute.manager [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 758.249377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.251036] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6931523c-847d-4721-8f01-164e9a1d36f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.263829] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.268406] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59d05d9c-5e8f-4d65-98b1-924302c358f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.270485] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453516, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.271852] env[65385]: WARNING neutronclient.v2_0.client [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.272577] env[65385]: WARNING openstack [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.272976] env[65385]: WARNING openstack [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.281444] env[65385]: DEBUG nova.compute.manager [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 758.281666] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.283719] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c28c84-4399-41bb-8ed0-9f1d2aff9346 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.294197] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.296715] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a84c3f32-2f8c-44ed-a668-dddbc0ea82a1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.298892] env[65385]: DEBUG oslo_vmware.api [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 758.298892] env[65385]: value = "task-4453520" [ 758.298892] env[65385]: _type = "Task" [ 758.298892] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.307454] env[65385]: DEBUG oslo_vmware.api [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 758.307454] env[65385]: value = "task-4453521" [ 758.307454] env[65385]: _type = "Task" [ 758.307454] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.314988] env[65385]: DEBUG oslo_vmware.api [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.327047] env[65385]: DEBUG oslo_vmware.api [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.330183] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453512, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651415} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.330456] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0/6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 758.330664] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.330934] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb3b3846-89d4-49e9-85d2-9e26d5974699 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.341539] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 758.341539] env[65385]: value = "task-4453522" [ 758.341539] env[65385]: _type = "Task" [ 758.341539] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.353537] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453522, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.450340] env[65385]: DEBUG nova.network.neutron [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] [instance: fae095a8-a768-4263-b456-ed0745398755] Updating instance_info_cache with network_info: [{"id": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "address": "fa:16:3e:9a:8f:26", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ca9c1b7-91", "ovs_interfaceid": "0ca9c1b7-9149-4367-af50-7af1b8ef0051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 758.496455] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.496798] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ed51dc6-6bc0-42bd-ada9-4d62164c2f81 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.507163] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 758.507163] env[65385]: value = "task-4453523" [ 758.507163] env[65385]: _type = "Task" [ 758.507163] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.519384] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453523, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.546685] env[65385]: DEBUG oslo_vmware.api [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Task: {'id': task-4453518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.384918} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.547603] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 758.547603] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 758.547968] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.547968] env[65385]: INFO nova.compute.manager [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Took 1.17 seconds to destroy the instance on the hypervisor. [ 758.548284] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 758.548456] env[65385]: DEBUG nova.compute.manager [-] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 758.548558] env[65385]: DEBUG nova.network.neutron [-] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 758.548931] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.549564] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.549867] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.597492] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.941s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.598056] env[65385]: DEBUG nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 758.603050] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.925s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.604387] env[65385]: INFO nova.compute.claims [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 758.646789] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.664273] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453519, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.749747] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453516, 'name': Destroy_Task, 'duration_secs': 0.785032} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.750032] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Destroyed the VM [ 758.750285] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 758.750660] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6d6bf0a2-4976-4ffd-be7c-09be4fb1c05d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.757926] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 758.757926] env[65385]: value = "task-4453524" [ 758.757926] env[65385]: _type = "Task" [ 758.757926] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.768275] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453524, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.815967] env[65385]: DEBUG oslo_vmware.api [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453520, 'name': PowerOffVM_Task, 'duration_secs': 0.368173} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.816825] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 758.817064] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 758.817546] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8145485d-3e92-421b-8c7b-324ca706decc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.823396] env[65385]: DEBUG oslo_vmware.api [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453521, 'name': PowerOffVM_Task, 'duration_secs': 0.392523} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.824083] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 758.824280] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 758.824608] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dcbf5313-36d9-44c9-abe4-61f66ed146e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.853720] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453522, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083587} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.854136] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.855145] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581d4efe-c558-430e-9d1a-4a6c55d192e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.884029] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0/6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.884029] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cddbb32-e3f8-49fe-8143-c0a35ed04f7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.906681] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 758.906681] env[65385]: value = "task-4453527" [ 758.906681] env[65385]: _type = "Task" [ 758.906681] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.918718] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453527, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.920527] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 758.920757] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 758.921691] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Deleting the datastore file [datastore2] fae095a8-a768-4263-b456-ed0745398755 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 758.921691] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-350be206-2041-418b-bc2c-b770bb30dc0d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.931315] env[65385]: DEBUG oslo_vmware.api [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for the task: (returnval){ [ 758.931315] env[65385]: value = "task-4453528" [ 758.931315] env[65385]: _type = "Task" [ 758.931315] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.932929] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 758.933152] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 758.933342] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Deleting the datastore file [datastore2] c798006c-3e74-461f-8397-3d8ca0be0a95 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 758.936985] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e8d27f5-f838-468d-b0b2-50bcbcf51e3f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.948628] env[65385]: DEBUG oslo_vmware.api [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.949548] env[65385]: DEBUG oslo_vmware.api [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 758.949548] env[65385]: value = "task-4453529" [ 758.949548] env[65385]: _type = "Task" [ 758.949548] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.953407] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bb5d853b-8175-4215-a1cd-598c31441bac tempest-ServerExternalEventsTest-1396571099 tempest-ServerExternalEventsTest-1396571099-project] Releasing lock "refresh_cache-fae095a8-a768-4263-b456-ed0745398755" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.960040] env[65385]: DEBUG oslo_vmware.api [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453529, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.020965] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453523, 'name': PowerOffVM_Task, 'duration_secs': 0.339178} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.021322] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 759.022633] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.023868] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2f68e3-8ac9-468a-99c5-10e201bc6467 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.032525] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 759.032927] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2bfae89-b596-41c3-980f-c628881652a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.072059] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c1cdca9c-74ad-4fb2-b66b-d923bd58253d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "7bdc813e-5e89-43e2-8963-502f26c998e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.072291] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c1cdca9c-74ad-4fb2-b66b-d923bd58253d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "7bdc813e-5e89-43e2-8963-502f26c998e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.075970] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 759.075970] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 759.076218] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Deleting the datastore file [datastore1] 53c31e42-4db0-4428-8336-b3a82e3f0e8f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.076516] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a37eb66-cf81-40e9-bc21-40554a556c40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.082134] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 759.082410] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 759.082599] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 759.082804] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 759.083055] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 759.083162] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 759.083397] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.083569] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 759.083750] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 759.083904] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 759.084128] env[65385]: DEBUG nova.virt.hardware [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 759.085581] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59dad147-afbe-4100-a419-4f2063b2e781 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.090609] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 759.090609] env[65385]: value = "task-4453531" [ 759.090609] env[65385]: _type = "Task" [ 759.090609] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.099060] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6309630c-1123-4cc7-b8ee-a805cc4ba5af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.107208] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.109263] env[65385]: DEBUG nova.compute.utils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 759.122588] env[65385]: DEBUG nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 759.122854] env[65385]: DEBUG nova.network.neutron [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 759.123312] env[65385]: WARNING neutronclient.v2_0.client [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.123672] env[65385]: WARNING neutronclient.v2_0.client [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.124428] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.124973] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.133393] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:0f:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6815237d-f565-474d-a3c0-9c675478eb00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2206d06f-505b-4211-9823-29d900820228', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.141284] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 759.143666] env[65385]: DEBUG nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 759.146614] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 759.148321] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f51e0fc-edd2-401e-abae-245b59f1b521 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.165996] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Received event network-vif-plugged-0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 759.166230] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Acquiring lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.166840] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.166840] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.166840] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] No waiting events found dispatching network-vif-plugged-0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 759.166840] env[65385]: WARNING nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Received unexpected event network-vif-plugged-0ee1cb97-07c8-4c45-9442-d76deacb6028 for instance with vm_state building and task_state spawning. [ 759.167168] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Received event network-changed-0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 759.167168] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Refreshing instance network info cache due to event network-changed-0ee1cb97-07c8-4c45-9442-d76deacb6028. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 759.168246] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Acquiring lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.168246] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Acquired lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.168246] env[65385]: DEBUG nova.network.neutron [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Refreshing network info cache for port 0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 759.182394] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453519, 'name': CreateVM_Task, 'duration_secs': 0.599845} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.187284] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 759.187750] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 759.187750] env[65385]: value = "task-4453532" [ 759.187750] env[65385]: _type = "Task" [ 759.187750] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.188114] env[65385]: WARNING neutronclient.v2_0.client [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.188579] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.188736] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.189074] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 759.189395] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f7e8bfe-326c-4f2f-a3a8-156e6244fc10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.200157] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 759.200157] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3a1b7-6f8b-310f-fa41-ae9e012926e2" [ 759.200157] env[65385]: _type = "Task" [ 759.200157] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.204066] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453532, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.214049] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3a1b7-6f8b-310f-fa41-ae9e012926e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.259137] env[65385]: DEBUG nova.policy [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '055d00bf06004425b894ab5ce8433a1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a80b4a1e9964a1da68fddeba39f07e3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 759.272626] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453524, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.419710] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.444578] env[65385]: DEBUG oslo_vmware.api [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Task: {'id': task-4453528, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256458} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.444848] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.445056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.445213] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.445482] env[65385]: INFO nova.compute.manager [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] [instance: fae095a8-a768-4263-b456-ed0745398755] Took 1.16 seconds to destroy the instance on the hypervisor. [ 759.445662] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 759.445879] env[65385]: DEBUG nova.compute.manager [-] [instance: fae095a8-a768-4263-b456-ed0745398755] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 759.445977] env[65385]: DEBUG nova.network.neutron [-] [instance: fae095a8-a768-4263-b456-ed0745398755] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 759.446311] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.447169] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.447291] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.468542] env[65385]: DEBUG oslo_vmware.api [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453529, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23476} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.469796] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.469796] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.469796] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.469796] env[65385]: INFO nova.compute.manager [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Took 1.22 seconds to destroy the instance on the hypervisor. [ 759.469796] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 759.470124] env[65385]: DEBUG nova.compute.manager [-] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 759.470124] env[65385]: DEBUG nova.network.neutron [-] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 759.470309] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.470918] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.471668] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.603998] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171021} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.604367] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.604526] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.604739] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.671549] env[65385]: INFO nova.virt.block_device [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Booting with volume 562ed51f-6604-4a40-be52-647787924294 at /dev/sda [ 759.677978] env[65385]: WARNING neutronclient.v2_0.client [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.678742] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.679137] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.714208] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453532, 'name': CreateVM_Task, 'duration_secs': 0.464644} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.718895] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 759.720628] env[65385]: WARNING neutronclient.v2_0.client [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 759.720628] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.720628] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.721263] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 759.721724] env[65385]: DEBUG nova.network.neutron [-] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 759.728169] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5910391-705f-4132-963f-443566a96702 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.729961] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3a1b7-6f8b-310f-fa41-ae9e012926e2, 'name': SearchDatastore_Task, 'duration_secs': 0.039764} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.731732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.731732] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 759.731732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.731903] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.732282] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 759.733297] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5be8e951-71fc-4654-ba6a-95e3ec36bf26 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.743115] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 759.743115] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52da0cdd-145c-5f35-dc34-ae561bde0a50" [ 759.743115] env[65385]: _type = "Task" [ 759.743115] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.754101] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b514323-96a1-465c-8285-244300051cea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.759200] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52da0cdd-145c-5f35-dc34-ae561bde0a50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.759482] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 759.759639] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 759.763383] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbc611d2-c84a-4bf5-bf90-f2c58da75172 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.773125] env[65385]: DEBUG nova.network.neutron [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Successfully created port: 94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 759.779666] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435b09f0-d3c9-420b-9924-248130c2d2da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.798321] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 759.798321] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ac6bcd-af94-3a63-a9dd-1d4353079511" [ 759.798321] env[65385]: _type = "Task" [ 759.798321] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.798648] env[65385]: DEBUG oslo_vmware.api [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453524, 'name': RemoveSnapshot_Task, 'duration_secs': 0.896021} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.799736] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 759.800516] env[65385]: INFO nova.compute.manager [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Took 17.29 seconds to snapshot the instance on the hypervisor. [ 759.813588] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ac6bcd-af94-3a63-a9dd-1d4353079511, 'name': SearchDatastore_Task, 'duration_secs': 0.027474} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.833301] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bc6c2d2-4c29-41fb-9b38-ce95cdb23ae4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.834476] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f574f5a-b407-4bcc-888e-14b0cf8f902b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.841361] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 759.841361] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5291d206-a81c-6a45-2a09-fc618de5be41" [ 759.841361] env[65385]: _type = "Task" [ 759.841361] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.849930] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c6d2b5-cc78-40e6-a94c-5f64e3e48cf7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.877208] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5291d206-a81c-6a45-2a09-fc618de5be41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.886348] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.907097] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473f591c-9603-47e0-96e9-e4f17c059747 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.916034] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.924546] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453527, 'name': ReconfigVM_Task, 'duration_secs': 0.909883} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.926762] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ad4645-4c0c-4bfb-8912-fe3e51802418 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.929547] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0/6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 759.929892] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=65385) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 759.930788] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-ff6e71b5-4404-4cb3-a0e5-f795853d2cea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.938823] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 759.938823] env[65385]: value = "task-4453533" [ 759.938823] env[65385]: _type = "Task" [ 759.938823] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.947944] env[65385]: DEBUG nova.virt.block_device [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Updating existing volume attachment record: 2576d1ca-6641-4efe-9c85-d9806087ba96 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 759.957920] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453533, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.231510] env[65385]: INFO nova.compute.manager [-] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Took 1.68 seconds to deallocate network for instance. [ 760.257907] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52da0cdd-145c-5f35-dc34-ae561bde0a50, 'name': SearchDatastore_Task, 'duration_secs': 0.019084} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.258271] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.258533] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.258793] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.258881] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 760.259072] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 760.259435] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74e7c6a8-8ce3-4c9b-afa8-40bf07e8a97b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.275943] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 760.276208] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 760.277175] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5729cf6b-17af-4b68-9eac-459a15c9d3e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.291219] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 760.291219] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5247bd17-87d0-6efa-d613-6c25c13655b8" [ 760.291219] env[65385]: _type = "Task" [ 760.291219] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.302908] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5247bd17-87d0-6efa-d613-6c25c13655b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.356169] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5291d206-a81c-6a45-2a09-fc618de5be41, 'name': SearchDatastore_Task, 'duration_secs': 0.040316} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.356651] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.356970] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 4232f3a4-c42b-4b9c-8e1d-da591634ffbb/4232f3a4-c42b-4b9c-8e1d-da591634ffbb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 760.357341] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e2b272a-91dc-4497-80eb-7c2015f69d05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.366671] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 760.366671] env[65385]: value = "task-4453534" [ 760.366671] env[65385]: _type = "Task" [ 760.366671] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.379072] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.388913] env[65385]: DEBUG nova.compute.manager [None req-a353d001-5022-45d9-8161-d83e7bae9760 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Found 1 images (rotation: 2) {{(pid=65385) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 760.455361] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453533, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.056172} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.456310] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=65385) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 760.457048] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67da063-7b31-420a-8ba7-3a0f1b6ee4a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.487597] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0/ephemeral_0.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.491150] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7211a22a-163f-4f45-8dc0-bea5faa3969e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.513644] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 760.513644] env[65385]: value = "task-4453535" [ 760.513644] env[65385]: _type = "Task" [ 760.513644] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.525260] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb524d7d-0835-4837-ba37-1c5cbe9fdab9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.532693] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453535, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.540835] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94727362-2da9-47f5-8b3c-daf3de062dc9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.580138] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6679345-1aba-4d07-ab7e-8d662b99c8dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.590543] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e59b76b-244e-4f63-ba2d-28e746ab08fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.611904] env[65385]: DEBUG nova.compute.provider_tree [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.654858] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 760.655095] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 760.655203] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 760.655411] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 760.655549] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 760.655689] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 760.655938] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.656092] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 760.656307] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 760.656528] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 760.656784] env[65385]: DEBUG nova.virt.hardware [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 760.658548] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d5060f-fbde-4310-aeed-daeebebfc3c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.670450] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99848ed2-ec52-48f9-a307-b8bea83f1e39 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.689629] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 760.697583] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 760.698537] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 760.698907] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1725d9b7-d3b6-4726-87c8-40e6e5ca7918 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.722015] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 760.722015] env[65385]: value = "task-4453536" [ 760.722015] env[65385]: _type = "Task" [ 760.722015] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.733303] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453536, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.739150] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 760.807421] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5247bd17-87d0-6efa-d613-6c25c13655b8, 'name': SearchDatastore_Task, 'duration_secs': 0.01484} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.808342] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc286636-4e7d-46ca-a98e-e063382900ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.816548] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 760.816548] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf7111-f748-978d-c54a-dc89d50e9651" [ 760.816548] env[65385]: _type = "Task" [ 760.816548] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.827595] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf7111-f748-978d-c54a-dc89d50e9651, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.879746] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453534, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.026795] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453535, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.048967] env[65385]: DEBUG nova.network.neutron [-] [instance: fae095a8-a768-4263-b456-ed0745398755] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 761.116220] env[65385]: DEBUG nova.scheduler.client.report [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.237020] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453536, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.335176] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf7111-f748-978d-c54a-dc89d50e9651, 'name': SearchDatastore_Task, 'duration_secs': 0.034831} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.335176] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 761.335555] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 761.335847] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73ae8016-a219-44ca-912a-f44cfe1d07ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.347036] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 761.347036] env[65385]: value = "task-4453537" [ 761.347036] env[65385]: _type = "Task" [ 761.347036] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.359971] env[65385]: DEBUG nova.network.neutron [-] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 761.361421] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453537, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.379318] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453534, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54061} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.380455] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 4232f3a4-c42b-4b9c-8e1d-da591634ffbb/4232f3a4-c42b-4b9c-8e1d-da591634ffbb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 761.381208] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 761.382365] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d92e0668-64ad-4d03-95df-d3cb2a522d59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.395134] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 761.395134] env[65385]: value = "task-4453538" [ 761.395134] env[65385]: _type = "Task" [ 761.395134] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.410204] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.526663] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453535, 'name': ReconfigVM_Task, 'duration_secs': 0.599007} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.526663] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0/ephemeral_0.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.527478] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30a4b8c0-a624-4539-8102-f06e55247075 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.537192] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 761.537192] env[65385]: value = "task-4453539" [ 761.537192] env[65385]: _type = "Task" [ 761.537192] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.548021] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453539, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.549257] env[65385]: DEBUG nova.network.neutron [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Successfully updated port: 94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 761.552298] env[65385]: INFO nova.compute.manager [-] [instance: fae095a8-a768-4263-b456-ed0745398755] Took 2.11 seconds to deallocate network for instance. [ 761.623033] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.021s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.624646] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 761.630459] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.633708] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.640314] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.920s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.643136] env[65385]: INFO nova.compute.claims [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.733206] env[65385]: WARNING neutronclient.v2_0.client [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 761.733206] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.733206] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.749582] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453536, 'name': CreateVM_Task, 'duration_secs': 0.539254} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.749864] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.750250] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.750525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 761.754120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 761.754120] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22b9da33-522d-4deb-b717-fa67783e598c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.759571] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 761.759571] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523b517f-0247-17e0-1bb7-6a597725b079" [ 761.759571] env[65385]: _type = "Task" [ 761.759571] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.769939] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523b517f-0247-17e0-1bb7-6a597725b079, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.863695] env[65385]: INFO nova.compute.manager [-] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Took 2.39 seconds to deallocate network for instance. [ 761.864074] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453537, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.910823] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075895} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.914813] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.914813] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc8429c-983f-41cb-b9ec-eb9f78d869ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.918084] env[65385]: DEBUG nova.network.neutron [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Updated VIF entry in instance network info cache for port 0ee1cb97-07c8-4c45-9442-d76deacb6028. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 761.918964] env[65385]: DEBUG nova.network.neutron [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Updating instance_info_cache with network_info: [{"id": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "address": "fa:16:3e:93:c1:c4", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ee1cb97-07", "ovs_interfaceid": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 761.951854] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 4232f3a4-c42b-4b9c-8e1d-da591634ffbb/4232f3a4-c42b-4b9c-8e1d-da591634ffbb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.952602] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bde255ad-8c3d-4fac-9a4b-862c94846569 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.979441] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 761.979441] env[65385]: value = "task-4453540" [ 761.979441] env[65385]: _type = "Task" [ 761.979441] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.989050] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453540, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.054987] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Acquiring lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.055323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Acquired lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.055375] env[65385]: DEBUG nova.network.neutron [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 762.056610] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453539, 'name': Rename_Task, 'duration_secs': 0.213667} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.056905] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.057219] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcafc8fb-a434-4721-a37a-ab28c71d089f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.061439] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.068789] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 762.068789] env[65385]: value = "task-4453541" [ 762.068789] env[65385]: _type = "Task" [ 762.068789] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.080790] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.096643] env[65385]: DEBUG nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 762.097442] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 762.097778] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 762.098212] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 762.098339] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 762.099779] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 762.099779] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 762.099779] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.099779] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 762.099779] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 762.099779] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 762.099779] env[65385]: DEBUG nova.virt.hardware [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 762.100547] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1a9a39-b9cb-49de-8328-66ebea09dc7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.112805] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2eb479-c28b-4d46-aca6-85e6c4927853 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.149393] env[65385]: DEBUG nova.compute.utils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 762.156365] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 762.156641] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 762.156984] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.157301] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.157912] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.158385] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.233046] env[65385]: DEBUG nova.policy [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38c0ae9a87cd47edb14f9ac8aa6cb3a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e412a22bd7b412996ee52adbea024bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 762.274404] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523b517f-0247-17e0-1bb7-6a597725b079, 'name': SearchDatastore_Task, 'duration_secs': 0.039594} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.274895] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.274895] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.275072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.275224] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.275405] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.275706] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f2d804e-84f5-4748-95b3-0b19adc550d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.289879] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.290106] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.291237] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e20572a-0435-482e-9047-6d9c96916e16 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.300209] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 762.300209] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b2d084-0221-5cf8-e096-ae9fcd635a39" [ 762.300209] env[65385]: _type = "Task" [ 762.300209] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.310946] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b2d084-0221-5cf8-e096-ae9fcd635a39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.360574] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453537, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543913} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.360930] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 762.361212] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.361567] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8be6d0a0-7f48-4236-90ce-1691dd0e30d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.372879] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 762.372879] env[65385]: value = "task-4453542" [ 762.372879] env[65385]: _type = "Task" [ 762.372879] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.380657] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.389448] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.421686] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Releasing lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.421876] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Received event network-vif-plugged-b74e7138-b571-449e-af1d-9e68bf8e9064 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 762.421975] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Acquiring lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.422345] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.422485] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.422600] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] No waiting events found dispatching network-vif-plugged-b74e7138-b571-449e-af1d-9e68bf8e9064 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 762.422769] env[65385]: WARNING nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Received unexpected event network-vif-plugged-b74e7138-b571-449e-af1d-9e68bf8e9064 for instance with vm_state building and task_state spawning. [ 762.422928] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Received event network-changed-b74e7138-b571-449e-af1d-9e68bf8e9064 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 762.423411] env[65385]: DEBUG nova.compute.manager [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Refreshing instance network info cache due to event network-changed-b74e7138-b571-449e-af1d-9e68bf8e9064. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 762.423411] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Acquiring lock "refresh_cache-4232f3a4-c42b-4b9c-8e1d-da591634ffbb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.423808] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Acquired lock "refresh_cache-4232f3a4-c42b-4b9c-8e1d-da591634ffbb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.424146] env[65385]: DEBUG nova.network.neutron [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Refreshing network info cache for port b74e7138-b571-449e-af1d-9e68bf8e9064 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 762.493474] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453540, 'name': ReconfigVM_Task, 'duration_secs': 0.370651} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.494029] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 4232f3a4-c42b-4b9c-8e1d-da591634ffbb/4232f3a4-c42b-4b9c-8e1d-da591634ffbb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 762.494658] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b8fc4a2-87bb-4cb8-ab2b-a7a40eb3fbb8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.503502] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 762.503502] env[65385]: value = "task-4453543" [ 762.503502] env[65385]: _type = "Task" [ 762.503502] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.515479] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453543, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.563377] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.563830] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.589086] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453541, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.633167] env[65385]: DEBUG nova.network.neutron [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 762.673421] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 762.754503] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.754939] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.770670] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Successfully created port: f18f6649-347b-43e6-9888-1dbb8af1227f {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 762.826960] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b2d084-0221-5cf8-e096-ae9fcd635a39, 'name': SearchDatastore_Task, 'duration_secs': 0.013029} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.830124] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-853a88e2-968b-4232-9ada-040e48b555c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.844195] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 762.844195] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529cf096-e191-0285-2f14-331ad138c51a" [ 762.844195] env[65385]: _type = "Task" [ 762.844195] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.859577] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529cf096-e191-0285-2f14-331ad138c51a, 'name': SearchDatastore_Task, 'duration_secs': 0.012287} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.859868] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.860286] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 762.860681] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f224e837-b290-4963-b5f5-cd6dfc5104b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.873084] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 762.873084] env[65385]: value = "task-4453544" [ 762.873084] env[65385]: _type = "Task" [ 762.873084] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.897648] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081559} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.897648] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.897648] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.898057] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0217b61-41bc-44cd-95fb-388d8f8fd260 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.928124] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.929828] env[65385]: WARNING neutronclient.v2_0.client [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.931215] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.931215] env[65385]: WARNING openstack [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.940417] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5995ebec-027e-4ee8-a3a4-bcfaae60a2a5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.959727] env[65385]: WARNING neutronclient.v2_0.client [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.961024] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.961024] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.969810] env[65385]: DEBUG nova.compute.manager [req-9520ec7b-6253-44c2-8015-7d8a7ad10447 req-fb555746-3e43-4e20-b32d-5e711dd4342d service nova] [instance: fae095a8-a768-4263-b456-ed0745398755] Received event network-vif-deleted-0ca9c1b7-9149-4367-af50-7af1b8ef0051 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 762.983130] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 762.983130] env[65385]: value = "task-4453545" [ 762.983130] env[65385]: _type = "Task" [ 762.983130] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.984128] env[65385]: DEBUG nova.compute.manager [req-7a9e0b9d-8791-49fc-a21c-5101789213f1 req-c5c18be8-cfb8-440b-887f-8e3a185da083 service nova] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Received event network-vif-deleted-d970616a-b8e1-4ab4-9f2e-e7b1f97dfff7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 763.000887] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453545, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.016416] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453543, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.087277] env[65385]: DEBUG oslo_vmware.api [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453541, 'name': PowerOnVM_Task, 'duration_secs': 0.567478} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.087277] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.087277] env[65385]: INFO nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Took 14.39 seconds to spawn the instance on the hypervisor. [ 763.087695] env[65385]: DEBUG nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 763.088455] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c467fa7-2fc0-4f65-a4e1-1fe81b8760bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.257115] env[65385]: DEBUG nova.network.neutron [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Updating instance_info_cache with network_info: [{"id": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "address": "fa:16:3e:d4:b5:c0", "network": {"id": "08330c64-94a6-41a4-b2a9-1641b6caaf43", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-741669912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a80b4a1e9964a1da68fddeba39f07e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ad8d4c-04", "ovs_interfaceid": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 763.275107] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.275536] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.386111] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453544, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.404167] env[65385]: WARNING neutronclient.v2_0.client [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.404904] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.405273] env[65385]: WARNING openstack [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.502779] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453545, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.516903] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453543, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.560889] env[65385]: DEBUG nova.network.neutron [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Updated VIF entry in instance network info cache for port b74e7138-b571-449e-af1d-9e68bf8e9064. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 763.561265] env[65385]: DEBUG nova.network.neutron [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Updating instance_info_cache with network_info: [{"id": "b74e7138-b571-449e-af1d-9e68bf8e9064", "address": "fa:16:3e:72:0c:9d", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.90", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb74e7138-b5", "ovs_interfaceid": "b74e7138-b571-449e-af1d-9e68bf8e9064", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 763.572144] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882bbfe5-0d11-46d8-8578-830185fb5ec6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.583047] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca933581-5edd-4b11-94c3-05dca754777b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.629511] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df90dde8-15cf-4619-a9da-6974b3f2a6ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.635235] env[65385]: INFO nova.compute.manager [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Took 51.85 seconds to build instance. [ 763.641447] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc8138e-1bb8-4696-ae87-69ba77c127fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.657617] env[65385]: DEBUG nova.compute.provider_tree [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.698132] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 763.721673] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 763.721673] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 763.721673] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 763.721673] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 763.721915] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 763.721915] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 763.722120] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.722276] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 763.722438] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 763.722593] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 763.722764] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 763.723696] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dba4524-e1da-4f07-8c16-c99448eb6e5d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.734524] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1b8e83-2b8f-4961-892e-9a361779b7c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.760312] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Releasing lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.760312] env[65385]: DEBUG nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Instance network_info: |[{"id": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "address": "fa:16:3e:d4:b5:c0", "network": {"id": "08330c64-94a6-41a4-b2a9-1641b6caaf43", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-741669912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a80b4a1e9964a1da68fddeba39f07e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ad8d4c-04", "ovs_interfaceid": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 763.760689] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:b5:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94ad8d4c-044d-465c-a0f7-b58c7f70cc14', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.768652] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Creating folder: Project (5a80b4a1e9964a1da68fddeba39f07e3). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 763.768979] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eefa91b8-aeb5-4dc0-af1a-a14e85d26a90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.783649] env[65385]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 763.783793] env[65385]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65385) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 763.784285] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Folder already exists: Project (5a80b4a1e9964a1da68fddeba39f07e3). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 763.785106] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Creating folder: Instances. Parent ref: group-v870975. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 763.785106] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f24c5288-3c0d-4cc0-91a2-2373052025ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.796631] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Created folder: Instances in parent group-v870975. [ 763.797101] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 763.797208] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 763.797721] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90cf2205-315d-435d-896d-5505be418b32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.818284] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.818284] env[65385]: value = "task-4453548" [ 763.818284] env[65385]: _type = "Task" [ 763.818284] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.830347] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453548, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.889140] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58071} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.889140] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 763.889140] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.889140] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ade45ba-bd7d-4b9a-8627-65a3b1aa24b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.899088] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 763.899088] env[65385]: value = "task-4453549" [ 763.899088] env[65385]: _type = "Task" [ 763.899088] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.912833] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.997342] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453545, 'name': ReconfigVM_Task, 'duration_secs': 0.632228} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.997342] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Reconfigured VM instance instance-00000014 to attach disk [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.998685] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b8e2432-908f-4e1a-b7f6-1b09993bb206 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.006539] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 764.006539] env[65385]: value = "task-4453550" [ 764.006539] env[65385]: _type = "Task" [ 764.006539] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.023223] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453543, 'name': Rename_Task, 'duration_secs': 1.192641} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.026841] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 764.027167] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453550, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.027419] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df4a9693-2d05-4cb5-91d9-6ccafa9ad125 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.039628] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 764.039628] env[65385]: value = "task-4453551" [ 764.039628] env[65385]: _type = "Task" [ 764.039628] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.050899] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453551, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.064787] env[65385]: DEBUG oslo_concurrency.lockutils [req-d554778a-7981-452a-bf71-c2b0a8f68d5f req-aac53faa-57b7-489f-9441-112ec4153c53 service nova] Releasing lock "refresh_cache-4232f3a4-c42b-4b9c-8e1d-da591634ffbb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.137789] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be080e72-375c-4b69-abed-2986bb066ef7 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.361s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.161604] env[65385]: DEBUG nova.scheduler.client.report [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.329945] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453548, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.413016] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081194} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.413587] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.414759] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f8291c-f3c0-4472-b072-f75e29f48690 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.441163] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.441163] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c0f8183-3e0c-46f7-9945-63f701ea09e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.463856] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 764.463856] env[65385]: value = "task-4453552" [ 764.463856] env[65385]: _type = "Task" [ 764.463856] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.475339] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453552, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.520586] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453550, 'name': Rename_Task, 'duration_secs': 0.166731} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.523710] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 764.523710] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f5a1948-48c2-417d-8afa-e9c4465c4c52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.530470] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 764.530470] env[65385]: value = "task-4453553" [ 764.530470] env[65385]: _type = "Task" [ 764.530470] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.544252] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.555666] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453551, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.580136] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 764.580136] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.614513] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Successfully updated port: f18f6649-347b-43e6-9888-1dbb8af1227f {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 764.640925] env[65385]: DEBUG nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 764.667690] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.028s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.669338] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 764.671728] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.772s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.673559] env[65385]: INFO nova.compute.claims [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.831368] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453548, 'name': CreateVM_Task, 'duration_secs': 0.699855} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.831819] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 764.835023] env[65385]: WARNING neutronclient.v2_0.client [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.835023] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870980', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'name': 'volume-562ed51f-6604-4a40-be52-647787924294', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1ad1e02b-fb82-4a88-9047-be316d590bd2', 'attached_at': '', 'detached_at': '', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'serial': '562ed51f-6604-4a40-be52-647787924294'}, 'device_type': None, 'guest_format': None, 'attachment_id': '2576d1ca-6641-4efe-9c85-d9806087ba96', 'mount_device': '/dev/sda', 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=65385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 764.835023] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Root volume attach. Driver type: vmdk {{(pid=65385) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 764.835023] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f91d15f-0f03-43b9-9b94-2142bc5b9f51 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.849893] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668cd9e6-3d58-4445-9074-be6d5e28b4d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.860731] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdd8ee4-928b-4b20-8b17-71b905624fa1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.869981] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-983c6895-c911-49d4-a875-7988a4f08eb1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.874103] env[65385]: DEBUG nova.compute.manager [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 764.874623] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf6dad9-b635-429e-8de7-ed80f729788d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.879373] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 764.879373] env[65385]: value = "task-4453554" [ 764.879373] env[65385]: _type = "Task" [ 764.879373] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.894355] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.975427] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453552, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.046490] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453553, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.060482] env[65385]: DEBUG oslo_vmware.api [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453551, 'name': PowerOnVM_Task, 'duration_secs': 0.542809} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.060884] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 765.061230] env[65385]: INFO nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Took 9.56 seconds to spawn the instance on the hypervisor. [ 765.061532] env[65385]: DEBUG nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 765.063066] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f548c0-75b9-4701-9eea-da5f81628d8a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.119286] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "refresh_cache-e3248f4a-8985-48b0-84cf-795f00ff21fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.119286] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "refresh_cache-e3248f4a-8985-48b0-84cf-795f00ff21fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.119286] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 765.168830] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.179756] env[65385]: DEBUG nova.compute.utils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 765.184765] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 765.185782] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 765.185782] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.186352] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.187415] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.187964] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.213550] env[65385]: DEBUG nova.compute.manager [req-76b21c85-cb2b-4a22-ba8c-bcc2ebcaf81f req-5ab64210-e6b4-4d75-969b-3574c18332e3 service nova] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Received event network-vif-deleted-ff2751e9-ab94-4f90-84f6-0d47f1514411 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 765.305563] env[65385]: DEBUG nova.policy [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38c0ae9a87cd47edb14f9ac8aa6cb3a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e412a22bd7b412996ee52adbea024bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 765.353240] env[65385]: DEBUG nova.compute.manager [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 765.353240] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e523894-5555-4e04-b2e6-6aa52fcc9381 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.398850] env[65385]: INFO nova.compute.manager [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] instance snapshotting [ 765.401598] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 35%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.402923] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba8e42a-8133-4ada-9561-9a3ec8ec95b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.431723] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1e0ad9-949e-45ba-bd0c-5be24760ebae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.478627] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453552, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.545209] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453553, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.587257] env[65385]: INFO nova.compute.manager [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Took 41.36 seconds to build instance. [ 765.624042] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.624813] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.704792] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 765.834810] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Successfully created port: c897de6b-a79c-436e-af74-6db933cc0c9f {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 765.868996] env[65385]: INFO nova.compute.manager [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] instance snapshotting [ 765.869889] env[65385]: DEBUG nova.objects.instance [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'flavor' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 765.899674] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 49%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.947085] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 765.951728] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-99fa61df-e049-4b53-a8ca-32bd0b454665 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.960814] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 765.960814] env[65385]: value = "task-4453555" [ 765.960814] env[65385]: _type = "Task" [ 765.960814] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.977901] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453555, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.986720] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453552, 'name': ReconfigVM_Task, 'duration_secs': 1.203617} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.987226] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 53c31e42-4db0-4428-8336-b3a82e3f0e8f/53c31e42-4db0-4428-8336-b3a82e3f0e8f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 765.989429] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9ffa8e7-a5d4-434e-962b-978b1b9cee26 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.004143] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 766.004143] env[65385]: value = "task-4453556" [ 766.004143] env[65385]: _type = "Task" [ 766.004143] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.021287] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 766.025131] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453556, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.034142] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Received event network-vif-plugged-94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 766.034305] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Acquiring lock "1ad1e02b-fb82-4a88-9047-be316d590bd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.034612] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.034879] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.035360] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] No waiting events found dispatching network-vif-plugged-94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 766.035527] env[65385]: WARNING nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Received unexpected event network-vif-plugged-94ad8d4c-044d-465c-a0f7-b58c7f70cc14 for instance with vm_state building and task_state spawning. [ 766.038120] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Received event network-changed-94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 766.038120] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Refreshing instance network info cache due to event network-changed-94ad8d4c-044d-465c-a0f7-b58c7f70cc14. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 766.038120] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Acquiring lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.038120] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Acquired lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.038120] env[65385]: DEBUG nova.network.neutron [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Refreshing network info cache for port 94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 766.055880] env[65385]: DEBUG oslo_vmware.api [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453553, 'name': PowerOnVM_Task, 'duration_secs': 1.300209} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.061204] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 766.061583] env[65385]: DEBUG nova.compute.manager [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 766.064507] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4aee85d-b524-4495-af32-2a5481d11c3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.093197] env[65385]: DEBUG oslo_concurrency.lockutils [None req-425784b1-a845-45c9-afef-9913d6c32e07 tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.152s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.106973] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.107615] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.334502] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.334502] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.334502] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.379757] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a205da32-13dc-4932-bd20-fc913db4393f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.411310] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47083392-ee4d-42ef-a261-b2e133d4ac56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.418514] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 62%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.481820] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453555, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.521468] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453556, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.546119] env[65385]: WARNING neutronclient.v2_0.client [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.547700] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.547700] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.569032] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Updating instance_info_cache with network_info: [{"id": "f18f6649-347b-43e6-9888-1dbb8af1227f", "address": "fa:16:3e:64:2d:ed", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf18f6649-34", "ovs_interfaceid": "f18f6649-347b-43e6-9888-1dbb8af1227f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.598353] env[65385]: DEBUG nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 766.608049] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.708565] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0752cb0a-87d8-4c15-8a3f-f1e43205b3b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.719767] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bacb0a0-00c5-454b-b08a-a30734843c68 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.725583] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 766.773799] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd94410b-2c29-49b4-a654-b714b30f5c1a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.786699] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 766.786976] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.787205] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 766.787438] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.787664] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 766.787811] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 766.788053] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.788251] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 766.788429] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 766.788594] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 766.788767] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 766.792450] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9398c4-1624-4cf5-be77-2f267bf66ba2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.795581] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be8f989-5c8f-4aa7-9e2e-9d95f31d14e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.811238] env[65385]: DEBUG nova.compute.provider_tree [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.814278] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.815156] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.828716] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7173eb5a-fe1f-49fa-b2eb-b73a78414c7b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.897411] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 73%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.929825] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 766.930331] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4f7c4a12-dc17-4e3e-a28a-ded875059f2e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.940453] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 766.940453] env[65385]: value = "task-4453557" [ 766.940453] env[65385]: _type = "Task" [ 766.940453] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.951352] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453557, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.952991] env[65385]: WARNING neutronclient.v2_0.client [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.953888] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.954375] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.970541] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.970541] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.983261] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453555, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.018905] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453556, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.071892] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "refresh_cache-e3248f4a-8985-48b0-84cf-795f00ff21fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.072191] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Instance network_info: |[{"id": "f18f6649-347b-43e6-9888-1dbb8af1227f", "address": "fa:16:3e:64:2d:ed", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf18f6649-34", "ovs_interfaceid": "f18f6649-347b-43e6-9888-1dbb8af1227f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 767.072698] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:2d:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f18f6649-347b-43e6-9888-1dbb8af1227f', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.081577] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Creating folder: Project (9e412a22bd7b412996ee52adbea024bf). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.081937] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f9a141b-4fd8-4a74-a992-2c57a9ed1153 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.097565] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Created folder: Project (9e412a22bd7b412996ee52adbea024bf) in parent group-v870881. [ 767.097780] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Creating folder: Instances. Parent ref: group-v871017. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.098271] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ee93eca-607c-4606-a045-72bbef71c9ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.117867] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Created folder: Instances in parent group-v871017. [ 767.117867] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 767.118031] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.118246] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a734b0ef-2dc1-4109-84b4-e6dfce80c09b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.136098] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.149020] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.149020] env[65385]: value = "task-4453560" [ 767.149020] env[65385]: _type = "Task" [ 767.149020] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.159208] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453560, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.212857] env[65385]: DEBUG nova.network.neutron [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Updated VIF entry in instance network info cache for port 94ad8d4c-044d-465c-a0f7-b58c7f70cc14. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 767.213392] env[65385]: DEBUG nova.network.neutron [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Updating instance_info_cache with network_info: [{"id": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "address": "fa:16:3e:d4:b5:c0", "network": {"id": "08330c64-94a6-41a4-b2a9-1641b6caaf43", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-741669912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a80b4a1e9964a1da68fddeba39f07e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ad8d4c-04", "ovs_interfaceid": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 767.327672] env[65385]: DEBUG nova.scheduler.client.report [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 767.348829] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.349249] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.401805] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 86%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.454370] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453557, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.483798] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453555, 'name': CreateSnapshot_Task, 'duration_secs': 1.348462} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.484159] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 767.485325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27308367-02b4-4a92-8073-8657e92668c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.518582] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453556, 'name': Rename_Task, 'duration_secs': 1.285439} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.520584] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 767.520979] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89b954b0-d2e8-47e8-b3f1-72028e587a73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.530971] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 767.530971] env[65385]: value = "task-4453561" [ 767.530971] env[65385]: _type = "Task" [ 767.530971] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.543798] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.666535] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453560, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.716870] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Releasing lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.717184] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Received event network-vif-plugged-f18f6649-347b-43e6-9888-1dbb8af1227f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 767.717416] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Acquiring lock "e3248f4a-8985-48b0-84cf-795f00ff21fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.717761] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.717961] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.717961] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] No waiting events found dispatching network-vif-plugged-f18f6649-347b-43e6-9888-1dbb8af1227f {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 767.718139] env[65385]: WARNING nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Received unexpected event network-vif-plugged-f18f6649-347b-43e6-9888-1dbb8af1227f for instance with vm_state building and task_state spawning. [ 767.718299] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Received event network-changed-f18f6649-347b-43e6-9888-1dbb8af1227f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 767.718655] env[65385]: DEBUG nova.compute.manager [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Refreshing instance network info cache due to event network-changed-f18f6649-347b-43e6-9888-1dbb8af1227f. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 767.718655] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Acquiring lock "refresh_cache-e3248f4a-8985-48b0-84cf-795f00ff21fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.718806] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Acquired lock "refresh_cache-e3248f4a-8985-48b0-84cf-795f00ff21fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.718987] env[65385]: DEBUG nova.network.neutron [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Refreshing network info cache for port f18f6649-347b-43e6-9888-1dbb8af1227f {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 767.821117] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Successfully updated port: c897de6b-a79c-436e-af74-6db933cc0c9f {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 767.836706] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.164s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.837167] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 767.841096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.865s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.842289] env[65385]: INFO nova.compute.claims [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.862613] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.863713] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.863713] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.863713] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.864071] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.864509] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.864509] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 767.864760] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.901238] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.948541] env[65385]: DEBUG nova.compute.manager [req-f283868f-e06d-40b0-ae71-ce771dada8bd req-fbad2515-2ebc-4621-acfc-50b462518c37 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Received event network-vif-plugged-c897de6b-a79c-436e-af74-6db933cc0c9f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 767.948662] env[65385]: DEBUG oslo_concurrency.lockutils [req-f283868f-e06d-40b0-ae71-ce771dada8bd req-fbad2515-2ebc-4621-acfc-50b462518c37 service nova] Acquiring lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.948893] env[65385]: DEBUG oslo_concurrency.lockutils [req-f283868f-e06d-40b0-ae71-ce771dada8bd req-fbad2515-2ebc-4621-acfc-50b462518c37 service nova] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.949082] env[65385]: DEBUG oslo_concurrency.lockutils [req-f283868f-e06d-40b0-ae71-ce771dada8bd req-fbad2515-2ebc-4621-acfc-50b462518c37 service nova] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.949260] env[65385]: DEBUG nova.compute.manager [req-f283868f-e06d-40b0-ae71-ce771dada8bd req-fbad2515-2ebc-4621-acfc-50b462518c37 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] No waiting events found dispatching network-vif-plugged-c897de6b-a79c-436e-af74-6db933cc0c9f {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 767.949428] env[65385]: WARNING nova.compute.manager [req-f283868f-e06d-40b0-ae71-ce771dada8bd req-fbad2515-2ebc-4621-acfc-50b462518c37 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Received unexpected event network-vif-plugged-c897de6b-a79c-436e-af74-6db933cc0c9f for instance with vm_state building and task_state spawning. [ 767.957026] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453557, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.011450] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 768.011807] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4c03b937-2840-45d5-9582-81e626d4acdf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.023038] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 768.023038] env[65385]: value = "task-4453562" [ 768.023038] env[65385]: _type = "Task" [ 768.023038] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.032903] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453562, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.042582] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453561, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.149056] env[65385]: DEBUG nova.compute.manager [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Received event network-changed-0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 768.149384] env[65385]: DEBUG nova.compute.manager [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Refreshing instance network info cache due to event network-changed-0ee1cb97-07c8-4c45-9442-d76deacb6028. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 768.149653] env[65385]: DEBUG oslo_concurrency.lockutils [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Acquiring lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.149842] env[65385]: DEBUG oslo_concurrency.lockutils [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Acquired lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.150154] env[65385]: DEBUG nova.network.neutron [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Refreshing network info cache for port 0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 768.164446] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453560, 'name': CreateVM_Task, 'duration_secs': 0.562296} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.165174] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.165691] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.167070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.167070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.167070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 768.167220] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f618d9a2-168b-4902-b3e7-5458f6732a31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.175105] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 768.175105] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e53354-8ce0-4003-3555-eb5b8502ad85" [ 768.175105] env[65385]: _type = "Task" [ 768.175105] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.187571] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e53354-8ce0-4003-3555-eb5b8502ad85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.228890] env[65385]: WARNING neutronclient.v2_0.client [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.228890] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.228890] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.247320] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5fdd988-935d-4165-8c1d-581b0baa5e9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.261681] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Suspending the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 768.261998] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-0c440640-e4ea-4f74-a52d-96a8afb40619 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.275224] env[65385]: DEBUG oslo_vmware.api [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] Waiting for the task: (returnval){ [ 768.275224] env[65385]: value = "task-4453563" [ 768.275224] env[65385]: _type = "Task" [ 768.275224] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.289246] env[65385]: DEBUG oslo_vmware.api [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] Task: {'id': task-4453563, 'name': SuspendVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.333849] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "refresh_cache-b702ba27-bbc7-4030-a96d-30e1ed656a9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.336248] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "refresh_cache-b702ba27-bbc7-4030-a96d-30e1ed656a9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.336570] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 768.352206] env[65385]: DEBUG nova.compute.utils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 768.355784] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 768.355784] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 768.356177] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.356614] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.357370] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.357779] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.370792] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.406729] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.454225] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453557, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.462773] env[65385]: DEBUG nova.policy [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38c0ae9a87cd47edb14f9ac8aa6cb3a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e412a22bd7b412996ee52adbea024bf', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 768.537947] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453562, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.549307] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453561, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.558736] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.559232] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.654151] env[65385]: WARNING neutronclient.v2_0.client [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.655102] env[65385]: WARNING openstack [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.655580] env[65385]: WARNING openstack [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.672665] env[65385]: WARNING neutronclient.v2_0.client [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.673370] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.673717] env[65385]: WARNING openstack [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.693922] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e53354-8ce0-4003-3555-eb5b8502ad85, 'name': SearchDatastore_Task, 'duration_secs': 0.03102} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.694312] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.694653] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.694943] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.695155] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.695356] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.695750] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4dee6fb9-740c-4f29-a3ac-07c4ba4ca0bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.709210] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.709454] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 768.713183] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d5396ab-6a1b-4700-88d2-2b22cc887c38 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.721093] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 768.721093] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9aee8-1080-8f1d-19c1-7d839ac6a423" [ 768.721093] env[65385]: _type = "Task" [ 768.721093] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.731482] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9aee8-1080-8f1d-19c1-7d839ac6a423, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.786850] env[65385]: DEBUG oslo_vmware.api [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] Task: {'id': task-4453563, 'name': SuspendVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.843869] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.844427] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.857339] env[65385]: INFO nova.compute.manager [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Rebuilding instance [ 768.859536] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 768.876179] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Successfully created port: 2034d6c4-ac55-40da-9bbe-a912b04d6548 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 768.914717] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task} progress is 98%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.923971] env[65385]: DEBUG nova.network.neutron [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Updated VIF entry in instance network info cache for port f18f6649-347b-43e6-9888-1dbb8af1227f. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 768.924896] env[65385]: DEBUG nova.network.neutron [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Updating instance_info_cache with network_info: [{"id": "f18f6649-347b-43e6-9888-1dbb8af1227f", "address": "fa:16:3e:64:2d:ed", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf18f6649-34", "ovs_interfaceid": "f18f6649-347b-43e6-9888-1dbb8af1227f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 768.931629] env[65385]: DEBUG nova.compute.manager [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 768.932682] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7e4c23-3e85-4f35-945f-3d346323a4b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.941098] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 768.962477] env[65385]: WARNING openstack [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.962843] env[65385]: WARNING openstack [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.974588] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453557, 'name': CreateSnapshot_Task, 'duration_secs': 1.624184} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.974893] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 768.975798] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980dea6a-2d22-4a3c-8c0c-c72a82b616c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.979926] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.980377] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.040029] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453562, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.060379] env[65385]: DEBUG oslo_vmware.api [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453561, 'name': PowerOnVM_Task, 'duration_secs': 1.106537} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.068479] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 769.068760] env[65385]: DEBUG nova.compute.manager [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 769.074322] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf63982-bbdc-4e34-a5ca-dc67fa4f4eab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.105306] env[65385]: WARNING neutronclient.v2_0.client [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.106036] env[65385]: WARNING openstack [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.106568] env[65385]: WARNING openstack [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.154788] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.155659] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.156018] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.233117] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9aee8-1080-8f1d-19c1-7d839ac6a423, 'name': SearchDatastore_Task, 'duration_secs': 0.016427} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.233976] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ea5e22a-6719-4b04-9fb4-cda0b9cebf10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.245836] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 769.245836] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52311a1b-33f0-861b-d728-ee1cde2f273b" [ 769.245836] env[65385]: _type = "Task" [ 769.245836] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.256116] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52311a1b-33f0-861b-d728-ee1cde2f273b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.293232] env[65385]: DEBUG oslo_vmware.api [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] Task: {'id': task-4453563, 'name': SuspendVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.373498] env[65385]: DEBUG nova.network.neutron [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Updated VIF entry in instance network info cache for port 0ee1cb97-07c8-4c45-9442-d76deacb6028. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 769.373913] env[65385]: DEBUG nova.network.neutron [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Updating instance_info_cache with network_info: [{"id": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "address": "fa:16:3e:93:c1:c4", "network": {"id": "56afd7ad-285b-45c3-8650-cf3dc89d5e60", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1382118133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2de8098800694299aae5aa2f59f405bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f68ebd2a-3c68-48db-8c32-8a01497fc2e7", "external-id": "nsx-vlan-transportzone-49", "segmentation_id": 49, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ee1cb97-07", "ovs_interfaceid": "0ee1cb97-07c8-4c45-9442-d76deacb6028", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.412277] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453554, 'name': RelocateVM_Task, 'duration_secs': 4.510023} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.412277] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 769.412277] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870980', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'name': 'volume-562ed51f-6604-4a40-be52-647787924294', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1ad1e02b-fb82-4a88-9047-be316d590bd2', 'attached_at': '', 'detached_at': '', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'serial': '562ed51f-6604-4a40-be52-647787924294'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 769.412461] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d8cc05-10a4-43c8-8bed-f6a8d041ff20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.432795] env[65385]: DEBUG oslo_concurrency.lockutils [req-eaad070d-eb9c-4448-aa55-3893a31ca3be req-665ef5a6-7edc-42c9-93d1-aa3d08c03713 service nova] Releasing lock "refresh_cache-e3248f4a-8985-48b0-84cf-795f00ff21fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.437229] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5a124f-ba62-46ae-9bd6-9c8618bf0d45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.465989] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] volume-562ed51f-6604-4a40-be52-647787924294/volume-562ed51f-6604-4a40-be52-647787924294.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.469477] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f261491d-7cfc-48e7-9bb0-7d637d4de701 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.488624] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Updating instance_info_cache with network_info: [{"id": "c897de6b-a79c-436e-af74-6db933cc0c9f", "address": "fa:16:3e:22:61:eb", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc897de6b-a7", "ovs_interfaceid": "c897de6b-a79c-436e-af74-6db933cc0c9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.509794] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 769.516264] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ff02687b-31ca-4064-9e93-6c1fd130dad0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.523026] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 769.523026] env[65385]: value = "task-4453564" [ 769.523026] env[65385]: _type = "Task" [ 769.523026] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.536917] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 769.536917] env[65385]: value = "task-4453565" [ 769.536917] env[65385]: _type = "Task" [ 769.536917] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.542642] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453564, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.560108] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453562, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.570305] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453565, 'name': CloneVM_Task} progress is 11%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.604469] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.712583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f363f4a-5034-49c1-900c-43926b929f05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.721770] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb34b3d-81ee-4130-920b-e3c8cacae84e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.759658] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0abadc-307d-4259-a681-487493c1eba1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.769162] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52311a1b-33f0-861b-d728-ee1cde2f273b, 'name': SearchDatastore_Task, 'duration_secs': 0.037426} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.771520] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.771791] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] e3248f4a-8985-48b0-84cf-795f00ff21fa/e3248f4a-8985-48b0-84cf-795f00ff21fa.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 769.772270] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-679e45b6-50b0-444a-a346-9ce189a78cff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.778106] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74009d3-b09c-4714-a666-e9b63618c90b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.785376] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 769.785376] env[65385]: value = "task-4453566" [ 769.785376] env[65385]: _type = "Task" [ 769.785376] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.801378] env[65385]: DEBUG nova.compute.provider_tree [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.803016] env[65385]: DEBUG oslo_vmware.api [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] Task: {'id': task-4453563, 'name': SuspendVM_Task, 'duration_secs': 1.078896} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.803635] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Suspended the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 769.803839] env[65385]: DEBUG nova.compute.manager [None req-a372d901-d90c-4ad2-af3e-51069a1950c3 tempest-ServersAdminNegativeTestJSON-266583657 tempest-ServersAdminNegativeTestJSON-266583657-project-admin] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 769.805591] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d962e8fd-1f4d-4ef1-b197-98437887752c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.812737] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453566, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.872071] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 769.876954] env[65385]: DEBUG oslo_concurrency.lockutils [req-53567704-125f-4ce9-b7e2-a66e2e57ab7e req-8ea7f600-dce5-4730-94a9-e0cccc5eebb3 service nova] Releasing lock "refresh_cache-6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.898394] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 769.898647] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 769.898835] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 769.898965] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 769.899137] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 769.899281] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 769.899487] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.899642] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 769.899804] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 769.899981] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 769.900191] env[65385]: DEBUG nova.virt.hardware [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 769.901453] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45359422-8109-4558-a2c8-6e4b35b1d2de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.913375] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd50643f-bf2d-4938-92b4-732e89aa09f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.997762] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "refresh_cache-b702ba27-bbc7-4030-a96d-30e1ed656a9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.998219] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Instance network_info: |[{"id": "c897de6b-a79c-436e-af74-6db933cc0c9f", "address": "fa:16:3e:22:61:eb", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc897de6b-a7", "ovs_interfaceid": "c897de6b-a79c-436e-af74-6db933cc0c9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 769.998925] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 769.999702] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:61:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c897de6b-a79c-436e-af74-6db933cc0c9f', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.008214] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 770.008451] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a25a894b-75d2-4962-9333-dcf643bd5c89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.010328] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 770.010640] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9510ae7-be29-41b3-8514-4abd9af470e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.037720] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 770.037720] env[65385]: value = "task-4453567" [ 770.037720] env[65385]: _type = "Task" [ 770.037720] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.039562] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 770.039562] env[65385]: value = "task-4453568" [ 770.039562] env[65385]: _type = "Task" [ 770.039562] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.047286] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453564, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.057754] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453562, 'name': CloneVM_Task, 'duration_secs': 1.706328} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.058777] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Created linked-clone VM from snapshot [ 770.060518] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892bb06a-39a2-421e-bfa8-7c34e28bdc76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.073780] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453567, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.074144] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453568, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.074505] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453565, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.081849] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Uploading image 3f16b034-e81d-4f27-a119-fc8ec7cb71a2 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 770.111495] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 770.111495] env[65385]: value = "vm-871021" [ 770.111495] env[65385]: _type = "VirtualMachine" [ 770.111495] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 770.111861] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f0d0c437-5e65-4538-b741-c2848f7c0e63 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.121042] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lease: (returnval){ [ 770.121042] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52755a77-d0af-71ff-f87b-eadfd42f40fc" [ 770.121042] env[65385]: _type = "HttpNfcLease" [ 770.121042] env[65385]: } obtained for exporting VM: (result){ [ 770.121042] env[65385]: value = "vm-871021" [ 770.121042] env[65385]: _type = "VirtualMachine" [ 770.121042] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 770.121656] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the lease: (returnval){ [ 770.121656] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52755a77-d0af-71ff-f87b-eadfd42f40fc" [ 770.121656] env[65385]: _type = "HttpNfcLease" [ 770.121656] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 770.131630] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 770.131630] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52755a77-d0af-71ff-f87b-eadfd42f40fc" [ 770.131630] env[65385]: _type = "HttpNfcLease" [ 770.131630] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 770.229669] env[65385]: DEBUG nova.compute.manager [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Received event network-changed-c897de6b-a79c-436e-af74-6db933cc0c9f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 770.229669] env[65385]: DEBUG nova.compute.manager [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Refreshing instance network info cache due to event network-changed-c897de6b-a79c-436e-af74-6db933cc0c9f. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 770.229837] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Acquiring lock "refresh_cache-b702ba27-bbc7-4030-a96d-30e1ed656a9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.230546] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Acquired lock "refresh_cache-b702ba27-bbc7-4030-a96d-30e1ed656a9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.230546] env[65385]: DEBUG nova.network.neutron [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Refreshing network info cache for port c897de6b-a79c-436e-af74-6db933cc0c9f {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 770.300308] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453566, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.310023] env[65385]: DEBUG nova.scheduler.client.report [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 770.354161] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.354740] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.355141] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.355434] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.355699] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.360261] env[65385]: INFO nova.compute.manager [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Terminating instance [ 770.509632] env[65385]: DEBUG nova.compute.manager [req-a9c08629-d11e-4720-b194-77c0bcbc204d req-f92a4d30-d467-4471-84b3-985550a6fc24 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Received event network-vif-plugged-2034d6c4-ac55-40da-9bbe-a912b04d6548 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 770.509875] env[65385]: DEBUG oslo_concurrency.lockutils [req-a9c08629-d11e-4720-b194-77c0bcbc204d req-f92a4d30-d467-4471-84b3-985550a6fc24 service nova] Acquiring lock "f521e37d-aa86-4533-9fbe-7d040ed90e53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.510212] env[65385]: DEBUG oslo_concurrency.lockutils [req-a9c08629-d11e-4720-b194-77c0bcbc204d req-f92a4d30-d467-4471-84b3-985550a6fc24 service nova] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.510435] env[65385]: DEBUG oslo_concurrency.lockutils [req-a9c08629-d11e-4720-b194-77c0bcbc204d req-f92a4d30-d467-4471-84b3-985550a6fc24 service nova] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.510776] env[65385]: DEBUG nova.compute.manager [req-a9c08629-d11e-4720-b194-77c0bcbc204d req-f92a4d30-d467-4471-84b3-985550a6fc24 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] No waiting events found dispatching network-vif-plugged-2034d6c4-ac55-40da-9bbe-a912b04d6548 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 770.511051] env[65385]: WARNING nova.compute.manager [req-a9c08629-d11e-4720-b194-77c0bcbc204d req-f92a4d30-d467-4471-84b3-985550a6fc24 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Received unexpected event network-vif-plugged-2034d6c4-ac55-40da-9bbe-a912b04d6548 for instance with vm_state building and task_state spawning. [ 770.548729] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453564, 'name': ReconfigVM_Task, 'duration_secs': 0.7585} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.558246] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Reconfigured VM instance instance-0000002e to attach disk [datastore2] volume-562ed51f-6604-4a40-be52-647787924294/volume-562ed51f-6604-4a40-be52-647787924294.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.567038] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35085afb-847e-43fc-b834-d115d620cd7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.583212] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453567, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.591376] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453565, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.597366] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453568, 'name': CreateVM_Task, 'duration_secs': 0.478018} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.597734] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 770.597734] env[65385]: value = "task-4453570" [ 770.597734] env[65385]: _type = "Task" [ 770.597734] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.597958] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 770.598560] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.598966] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.599220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.599503] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 770.603155] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-853a5bd6-ace9-413a-be78-14fe438feb8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.613821] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.615799] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 770.615799] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5227cf51-9ab2-2448-dd8f-0b270e43276f" [ 770.615799] env[65385]: _type = "Task" [ 770.615799] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.628985] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5227cf51-9ab2-2448-dd8f-0b270e43276f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.635099] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 770.635099] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52755a77-d0af-71ff-f87b-eadfd42f40fc" [ 770.635099] env[65385]: _type = "HttpNfcLease" [ 770.635099] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 770.635099] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 770.635099] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52755a77-d0af-71ff-f87b-eadfd42f40fc" [ 770.635099] env[65385]: _type = "HttpNfcLease" [ 770.635099] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 770.635099] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171dab22-ca73-424c-9522-b0e536145e6c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.640294] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Successfully updated port: 2034d6c4-ac55-40da-9bbe-a912b04d6548 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 770.649760] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c362a-92de-47ee-9954-186aef65e23e/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 770.651417] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c362a-92de-47ee-9954-186aef65e23e/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 770.733413] env[65385]: WARNING neutronclient.v2_0.client [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.734178] env[65385]: WARNING openstack [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.734650] env[65385]: WARNING openstack [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.746654] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-14213f07-3701-4dcd-a139-cdd642f1f1a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.797871] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453566, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.899085} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.798143] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] e3248f4a-8985-48b0-84cf-795f00ff21fa/e3248f4a-8985-48b0-84cf-795f00ff21fa.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 770.798347] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.798630] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9443ddfd-bd56-4946-bed5-9281debbc41c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.807257] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 770.807257] env[65385]: value = "task-4453571" [ 770.807257] env[65385]: _type = "Task" [ 770.807257] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.818126] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.977s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.818714] env[65385]: DEBUG nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 770.821385] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.821658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.609s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.823133] env[65385]: INFO nova.compute.claims [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.866915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "refresh_cache-53c31e42-4db0-4428-8336-b3a82e3f0e8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.866915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquired lock "refresh_cache-53c31e42-4db0-4428-8336-b3a82e3f0e8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.867146] env[65385]: DEBUG nova.network.neutron [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 770.880145] env[65385]: WARNING openstack [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.880669] env[65385]: WARNING openstack [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.951447] env[65385]: WARNING neutronclient.v2_0.client [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.952578] env[65385]: WARNING openstack [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.952895] env[65385]: WARNING openstack [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.067774] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453567, 'name': PowerOffVM_Task, 'duration_secs': 0.524259} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.074678] env[65385]: DEBUG nova.network.neutron [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Updated VIF entry in instance network info cache for port c897de6b-a79c-436e-af74-6db933cc0c9f. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 771.075320] env[65385]: DEBUG nova.network.neutron [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Updating instance_info_cache with network_info: [{"id": "c897de6b-a79c-436e-af74-6db933cc0c9f", "address": "fa:16:3e:22:61:eb", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc897de6b-a7", "ovs_interfaceid": "c897de6b-a79c-436e-af74-6db933cc0c9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.076790] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 771.080021] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.080021] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453565, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.080021] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1d10aa-5fed-461e-b429-01ac20f1b1a5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.088947] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 771.090287] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5c22f99-e9ad-4f97-974a-07e1f3b3442f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.113474] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.128881] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5227cf51-9ab2-2448-dd8f-0b270e43276f, 'name': SearchDatastore_Task, 'duration_secs': 0.053731} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.129336] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.129697] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 771.130015] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.130235] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.130644] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 771.131540] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05c6ff8b-6673-48f9-a74d-860edd6dcf59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.142438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "refresh_cache-f521e37d-aa86-4533-9fbe-7d040ed90e53" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.142727] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "refresh_cache-f521e37d-aa86-4533-9fbe-7d040ed90e53" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.143089] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 771.144698] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 771.145057] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 771.146527] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e92e534f-0c63-4da6-bcbd-af6d8413f597 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.160247] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 771.160247] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52de963e-265b-d635-dbaa-843e863c460d" [ 771.160247] env[65385]: _type = "Task" [ 771.160247] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.174619] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52de963e-265b-d635-dbaa-843e863c460d, 'name': SearchDatastore_Task, 'duration_secs': 0.013219} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.175419] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-413fff6d-4d83-466b-af89-e27c02c9efb8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.184532] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 771.184532] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f6005e-1e1a-ef5a-8311-3d8f1ba11b8b" [ 771.184532] env[65385]: _type = "Task" [ 771.184532] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.199015] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f6005e-1e1a-ef5a-8311-3d8f1ba11b8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.201139] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 771.202469] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 771.202680] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleting the datastore file [datastore1] eb9d0510-b453-4695-9e1d-731217b9f8ec {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.203080] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97e33f02-96ab-48dc-b248-6c21d6517411 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.214306] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 771.214306] env[65385]: value = "task-4453573" [ 771.214306] env[65385]: _type = "Task" [ 771.214306] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.226365] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.322522] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070635} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.323234] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.324314] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe30d66a-b016-4512-9fb8-6f04d8c3fdf8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.329401] env[65385]: DEBUG nova.compute.utils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 771.333921] env[65385]: DEBUG nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 771.335283] env[65385]: DEBUG nova.network.neutron [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 771.335283] env[65385]: WARNING neutronclient.v2_0.client [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.335283] env[65385]: WARNING neutronclient.v2_0.client [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.335656] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.335908] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.369214] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] e3248f4a-8985-48b0-84cf-795f00ff21fa/e3248f4a-8985-48b0-84cf-795f00ff21fa.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.370525] env[65385]: WARNING neutronclient.v2_0.client [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.371286] env[65385]: WARNING openstack [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.371935] env[65385]: WARNING openstack [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.379652] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3436ef0-2a9d-4b1e-a7d7-740daa09f49b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.404167] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 771.404167] env[65385]: value = "task-4453574" [ 771.404167] env[65385]: _type = "Task" [ 771.404167] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.415101] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.419095] env[65385]: DEBUG nova.policy [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bbc524f6b2e64dd0af9d06525190299e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99f640b7d5594400bbd69c05713089d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 771.430992] env[65385]: DEBUG nova.network.neutron [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 771.543356] env[65385]: DEBUG nova.network.neutron [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.560269] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453565, 'name': CloneVM_Task, 'duration_secs': 1.869452} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.560512] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Created linked-clone VM from snapshot [ 771.561740] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ebe864-4fbc-496b-8e54-2eff6cb1bd58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.576388] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Uploading image 6597b3e2-a8ce-4ef5-81d7-597aba2aa3d6 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 771.580700] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5272e05-8448-4725-a4bf-0464751dd1a8 req-2dbc3a42-5c6b-415f-bc15-2d357e60fb09 service nova] Releasing lock "refresh_cache-b702ba27-bbc7-4030-a96d-30e1ed656a9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.608026] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 771.608026] env[65385]: value = "vm-871023" [ 771.608026] env[65385]: _type = "VirtualMachine" [ 771.608026] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 771.609239] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c9ecc03d-d7f7-4a96-ab23-9d9fecff2a7b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.621370] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453570, 'name': ReconfigVM_Task, 'duration_secs': 1.028304} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.623073] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870980', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'name': 'volume-562ed51f-6604-4a40-be52-647787924294', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1ad1e02b-fb82-4a88-9047-be316d590bd2', 'attached_at': '', 'detached_at': '', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'serial': '562ed51f-6604-4a40-be52-647787924294'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 771.624034] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease: (returnval){ [ 771.624034] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c5d5c-b91d-b7df-dda9-93179d4a8cdd" [ 771.624034] env[65385]: _type = "HttpNfcLease" [ 771.624034] env[65385]: } obtained for exporting VM: (result){ [ 771.624034] env[65385]: value = "vm-871023" [ 771.624034] env[65385]: _type = "VirtualMachine" [ 771.624034] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 771.625557] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the lease: (returnval){ [ 771.625557] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c5d5c-b91d-b7df-dda9-93179d4a8cdd" [ 771.625557] env[65385]: _type = "HttpNfcLease" [ 771.625557] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 771.625557] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc85794a-190f-45e1-ac80-fe7673803916 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.637904] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 771.637904] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c5d5c-b91d-b7df-dda9-93179d4a8cdd" [ 771.637904] env[65385]: _type = "HttpNfcLease" [ 771.637904] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 771.641023] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 771.641023] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c5d5c-b91d-b7df-dda9-93179d4a8cdd" [ 771.641023] env[65385]: _type = "HttpNfcLease" [ 771.641023] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 771.641023] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 771.641023] env[65385]: value = "task-4453576" [ 771.641023] env[65385]: _type = "Task" [ 771.641023] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.641023] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a34fea-88a9-4ca1-ae19-a33fc23a4f2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.648190] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.648746] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.662399] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a7a1c1-d07d-5f5b-ca3d-22a8016137e1/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 771.662673] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a7a1c1-d07d-5f5b-ca3d-22a8016137e1/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 771.669062] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453576, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.742102] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 771.768998] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f6005e-1e1a-ef5a-8311-3d8f1ba11b8b, 'name': SearchDatastore_Task, 'duration_secs': 0.024468} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.774799] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.775418] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] b702ba27-bbc7-4030-a96d-30e1ed656a9b/b702ba27-bbc7-4030-a96d-30e1ed656a9b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 771.776705] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238813} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.776705] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd5b7900-2d81-4568-9578-f31335ddf994 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.780049] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.780454] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 771.780850] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.786727] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.787369] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.806578] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 771.806578] env[65385]: value = "task-4453577" [ 771.806578] env[65385]: _type = "Task" [ 771.806578] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.819236] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.829990] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e100b21a-f98e-4274-830b-c006ecc804f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.838622] env[65385]: DEBUG nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 771.856019] env[65385]: DEBUG nova.network.neutron [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Successfully created port: a150e5dc-1294-45e5-9346-d31c443b7efe {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 771.918303] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453574, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.961584] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.963122] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.964956] env[65385]: WARNING openstack [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.050824] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Releasing lock "refresh_cache-53c31e42-4db0-4428-8336-b3a82e3f0e8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.051767] env[65385]: DEBUG nova.compute.manager [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 772.052342] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 772.053455] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdae0e5-801b-4524-8ae1-4dd880935116 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.068311] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 772.068586] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2c49088-bcdc-4c29-b004-8d63f673ec3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.079022] env[65385]: DEBUG oslo_vmware.api [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 772.079022] env[65385]: value = "task-4453578" [ 772.079022] env[65385]: _type = "Task" [ 772.079022] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.097805] env[65385]: DEBUG oslo_vmware.api [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.157988] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453576, 'name': Rename_Task, 'duration_secs': 0.23957} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.160144] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.160144] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8af72d5f-5ef0-4a24-8adc-4f835087faf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.173203] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 772.173203] env[65385]: value = "task-4453579" [ 772.173203] env[65385]: _type = "Task" [ 772.173203] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.190156] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.193276] env[65385]: DEBUG nova.network.neutron [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Updating instance_info_cache with network_info: [{"id": "2034d6c4-ac55-40da-9bbe-a912b04d6548", "address": "fa:16:3e:8e:64:f6", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2034d6c4-ac", "ovs_interfaceid": "2034d6c4-ac55-40da-9bbe-a912b04d6548", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 772.328634] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.420234] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453574, 'name': ReconfigVM_Task, 'duration_secs': 0.642014} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.420463] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Reconfigured VM instance instance-0000002f to attach disk [datastore1] e3248f4a-8985-48b0-84cf-795f00ff21fa/e3248f4a-8985-48b0-84cf-795f00ff21fa.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.423342] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52d186b3-90b4-4d95-83ef-eb8ad7b6137f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.430620] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 772.430620] env[65385]: value = "task-4453580" [ 772.430620] env[65385]: _type = "Task" [ 772.430620] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.449239] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453580, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.593910] env[65385]: DEBUG oslo_vmware.api [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.663393] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322b3659-41d7-49d3-a3ea-29400ca78cbd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.673147] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68230cb6-56e6-4203-90bc-10cede5de156 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.721443] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "refresh_cache-f521e37d-aa86-4533-9fbe-7d040ed90e53" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.721716] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Instance network_info: |[{"id": "2034d6c4-ac55-40da-9bbe-a912b04d6548", "address": "fa:16:3e:8e:64:f6", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2034d6c4-ac", "ovs_interfaceid": "2034d6c4-ac55-40da-9bbe-a912b04d6548", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 772.727017] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:64:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2034d6c4-ac55-40da-9bbe-a912b04d6548', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.737603] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 772.738512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ef2197-9187-4b20-a945-a32047435a99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.742263] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453579, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.742641] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 772.742998] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfa2c158-f441-4d68-9457-cde50fc436a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.767937] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd6c9d2-01fc-4e0b-8bd7-65f9830cf62f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.774498] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.774498] env[65385]: value = "task-4453581" [ 772.774498] env[65385]: _type = "Task" [ 772.774498] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.789395] env[65385]: DEBUG nova.compute.provider_tree [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.797598] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453581, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.826237] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453577, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.851869] env[65385]: DEBUG nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 772.950873] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453580, 'name': Rename_Task, 'duration_secs': 0.215301} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.951482] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.951911] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c812e35-819c-48a3-b64b-16b3626240d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.961886] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 772.961886] env[65385]: value = "task-4453582" [ 772.961886] env[65385]: _type = "Task" [ 772.961886] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.974120] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.094962] env[65385]: DEBUG oslo_vmware.api [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453578, 'name': PowerOffVM_Task, 'duration_secs': 0.772123} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.095671] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.096056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.096649] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f21c329-07fc-41d3-b45a-899cc8d9a456 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.140906] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.141335] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.141550] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Deleting the datastore file [datastore2] 53c31e42-4db0-4428-8336-b3a82e3f0e8f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.141974] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4208579-6350-4e8e-9a17-381593186f33 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.152563] env[65385]: DEBUG oslo_vmware.api [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for the task: (returnval){ [ 773.152563] env[65385]: value = "task-4453584" [ 773.152563] env[65385]: _type = "Task" [ 773.152563] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.164975] env[65385]: DEBUG oslo_vmware.api [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453584, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.191605] env[65385]: DEBUG oslo_vmware.api [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453579, 'name': PowerOnVM_Task, 'duration_secs': 0.549842} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.192284] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.192722] env[65385]: INFO nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Took 11.10 seconds to spawn the instance on the hypervisor. [ 773.193153] env[65385]: DEBUG nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 773.194415] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9895f789-79cc-499d-9b41-5e9adf12c54c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.254548] env[65385]: DEBUG nova.compute.manager [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Received event network-changed-2034d6c4-ac55-40da-9bbe-a912b04d6548 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 773.255097] env[65385]: DEBUG nova.compute.manager [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Refreshing instance network info cache due to event network-changed-2034d6c4-ac55-40da-9bbe-a912b04d6548. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 773.255248] env[65385]: DEBUG oslo_concurrency.lockutils [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Acquiring lock "refresh_cache-f521e37d-aa86-4533-9fbe-7d040ed90e53" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.255337] env[65385]: DEBUG oslo_concurrency.lockutils [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Acquired lock "refresh_cache-f521e37d-aa86-4533-9fbe-7d040ed90e53" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.255781] env[65385]: DEBUG nova.network.neutron [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Refreshing network info cache for port 2034d6c4-ac55-40da-9bbe-a912b04d6548 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 773.289829] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453581, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.294309] env[65385]: DEBUG nova.scheduler.client.report [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 773.326563] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453577, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.514463} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.327119] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] b702ba27-bbc7-4030-a96d-30e1ed656a9b/b702ba27-bbc7-4030-a96d-30e1ed656a9b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 773.327573] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 773.328442] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-617c0a0b-e211-4768-aa24-3609767357ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.338045] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 773.338045] env[65385]: value = "task-4453585" [ 773.338045] env[65385]: _type = "Task" [ 773.338045] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.351534] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453585, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.475939] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453582, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.624258] env[65385]: DEBUG nova.network.neutron [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Successfully updated port: a150e5dc-1294-45e5-9346-d31c443b7efe {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 773.666373] env[65385]: DEBUG oslo_vmware.api [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Task: {'id': task-4453584, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280389} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.667441] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.667654] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 773.669057] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 773.669057] env[65385]: INFO nova.compute.manager [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Took 1.62 seconds to destroy the instance on the hypervisor. [ 773.669057] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 773.669273] env[65385]: DEBUG nova.compute.manager [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 773.669273] env[65385]: DEBUG nova.network.neutron [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 773.669511] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.670304] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.670676] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.696953] env[65385]: DEBUG nova.network.neutron [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 773.697445] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.720045] env[65385]: INFO nova.compute.manager [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Took 47.67 seconds to build instance. [ 773.760347] env[65385]: WARNING neutronclient.v2_0.client [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.761142] env[65385]: WARNING openstack [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.761617] env[65385]: WARNING openstack [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.787460] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453581, 'name': CreateVM_Task, 'duration_secs': 0.552958} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.787792] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 773.788448] env[65385]: WARNING neutronclient.v2_0.client [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.788863] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.789038] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.789472] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 773.790216] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9407a4d-c01d-49d1-94a5-82fcf664e12d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.796758] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 773.796758] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527a43e9-8c24-bd2d-0ea2-9232fa8a37cf" [ 773.796758] env[65385]: _type = "Task" [ 773.796758] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.801770] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.980s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.802617] env[65385]: DEBUG nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 773.805792] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.120s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.806201] env[65385]: DEBUG nova.objects.instance [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lazy-loading 'resources' on Instance uuid 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.814858] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527a43e9-8c24-bd2d-0ea2-9232fa8a37cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.850748] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453585, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093278} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.851150] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 773.852324] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bb0bd7-3e17-4c86-a068-ff9a1eb6ae0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.885310] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] b702ba27-bbc7-4030-a96d-30e1ed656a9b/b702ba27-bbc7-4030-a96d-30e1ed656a9b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 773.892462] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1919244e-3fbf-4627-a27c-efedd776e819 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.921687] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 773.921687] env[65385]: value = "task-4453586" [ 773.921687] env[65385]: _type = "Task" [ 773.921687] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.933715] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453586, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.938416] env[65385]: WARNING openstack [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.939116] env[65385]: WARNING openstack [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.975988] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453582, 'name': PowerOnVM_Task, 'duration_secs': 0.699305} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.981340] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.981672] env[65385]: INFO nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Took 10.28 seconds to spawn the instance on the hypervisor. [ 773.981986] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 773.983199] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9ea01a-e1a3-4a9f-89c8-9ba47f13bc0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.016030] env[65385]: WARNING neutronclient.v2_0.client [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.016426] env[65385]: WARNING openstack [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.016934] env[65385]: WARNING openstack [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.115865] env[65385]: DEBUG nova.network.neutron [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Updated VIF entry in instance network info cache for port 2034d6c4-ac55-40da-9bbe-a912b04d6548. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 774.116337] env[65385]: DEBUG nova.network.neutron [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Updating instance_info_cache with network_info: [{"id": "2034d6c4-ac55-40da-9bbe-a912b04d6548", "address": "fa:16:3e:8e:64:f6", "network": {"id": "c8e298a3-44d1-4599-8e05-7336cfa8e2e8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1352334807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e412a22bd7b412996ee52adbea024bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2034d6c4-ac", "ovs_interfaceid": "2034d6c4-ac55-40da-9bbe-a912b04d6548", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 774.127957] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "refresh_cache-98430d7e-626e-4532-802e-38f7fc583ae3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.128241] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "refresh_cache-98430d7e-626e-4532-802e-38f7fc583ae3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 774.128428] env[65385]: DEBUG nova.network.neutron [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 774.200655] env[65385]: DEBUG nova.network.neutron [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 774.223104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c86af85-c43b-4538-a725-664e17ebb69e tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.534s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.247028] env[65385]: DEBUG nova.compute.manager [req-cb70419c-100a-4b54-a53f-f3e6931d6ff3 req-99d2813b-dec9-49e9-9894-e05b9b5535d1 service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Received event network-vif-plugged-a150e5dc-1294-45e5-9346-d31c443b7efe {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 774.247389] env[65385]: DEBUG oslo_concurrency.lockutils [req-cb70419c-100a-4b54-a53f-f3e6931d6ff3 req-99d2813b-dec9-49e9-9894-e05b9b5535d1 service nova] Acquiring lock "98430d7e-626e-4532-802e-38f7fc583ae3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 774.247694] env[65385]: DEBUG oslo_concurrency.lockutils [req-cb70419c-100a-4b54-a53f-f3e6931d6ff3 req-99d2813b-dec9-49e9-9894-e05b9b5535d1 service nova] Lock "98430d7e-626e-4532-802e-38f7fc583ae3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.247921] env[65385]: DEBUG oslo_concurrency.lockutils [req-cb70419c-100a-4b54-a53f-f3e6931d6ff3 req-99d2813b-dec9-49e9-9894-e05b9b5535d1 service nova] Lock "98430d7e-626e-4532-802e-38f7fc583ae3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.248125] env[65385]: DEBUG nova.compute.manager [req-cb70419c-100a-4b54-a53f-f3e6931d6ff3 req-99d2813b-dec9-49e9-9894-e05b9b5535d1 service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] No waiting events found dispatching network-vif-plugged-a150e5dc-1294-45e5-9346-d31c443b7efe {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 774.248315] env[65385]: WARNING nova.compute.manager [req-cb70419c-100a-4b54-a53f-f3e6931d6ff3 req-99d2813b-dec9-49e9-9894-e05b9b5535d1 service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Received unexpected event network-vif-plugged-a150e5dc-1294-45e5-9346-d31c443b7efe for instance with vm_state building and task_state spawning. [ 774.314320] env[65385]: DEBUG nova.compute.utils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 774.320880] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527a43e9-8c24-bd2d-0ea2-9232fa8a37cf, 'name': SearchDatastore_Task, 'duration_secs': 0.017835} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.322881] env[65385]: DEBUG nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 774.322881] env[65385]: DEBUG nova.network.neutron [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 774.323706] env[65385]: WARNING neutronclient.v2_0.client [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.324189] env[65385]: WARNING neutronclient.v2_0.client [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.325115] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.325776] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.337208] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 774.337705] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 774.338252] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.338572] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 774.338836] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 774.343243] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfb062f9-b828-42d0-a311-fcb1ba7d9f76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.355341] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 774.356724] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 774.358536] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebc4c1a8-e522-418d-b8a8-afe08bf5ddab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.368587] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 774.368587] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52885eaf-51fe-e8e1-c8d1-8d146a505dc6" [ 774.368587] env[65385]: _type = "Task" [ 774.368587] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.380591] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52885eaf-51fe-e8e1-c8d1-8d146a505dc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.416368] env[65385]: DEBUG nova.policy [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0f069272314a31ad33c77137513a17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c27eb6232a2148a1a259f57494b4ae30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 774.432976] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453586, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.506018] env[65385]: INFO nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Took 42.86 seconds to build instance. [ 774.619175] env[65385]: DEBUG oslo_concurrency.lockutils [req-174611ee-2a18-4496-bd60-be3f2d7f8275 req-463a4e44-056c-4d54-82bb-c7815cc7c572 service nova] Releasing lock "refresh_cache-f521e37d-aa86-4533-9fbe-7d040ed90e53" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 774.633483] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.634111] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.684563] env[65385]: DEBUG nova.network.neutron [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 774.703478] env[65385]: INFO nova.compute.manager [-] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Took 1.03 seconds to deallocate network for instance. [ 774.724060] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.724782] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.732412] env[65385]: DEBUG nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 774.802611] env[65385]: WARNING neutronclient.v2_0.client [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.803356] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.803748] env[65385]: WARNING openstack [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.816946] env[65385]: DEBUG nova.network.neutron [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Successfully created port: 3617c8a9-2863-4483-86be-36b932f830a3 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 774.822032] env[65385]: DEBUG nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 774.886413] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52885eaf-51fe-e8e1-c8d1-8d146a505dc6, 'name': SearchDatastore_Task, 'duration_secs': 0.015352} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.888983] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d4e0d4a-3ef7-48de-aab0-d652763ab00c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.896694] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 774.896694] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3bc1e-59d9-3e61-9c65-85a3ec5eff39" [ 774.896694] env[65385]: _type = "Task" [ 774.896694] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.902548] env[65385]: DEBUG nova.network.neutron [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Updating instance_info_cache with network_info: [{"id": "a150e5dc-1294-45e5-9346-d31c443b7efe", "address": "fa:16:3e:c8:5a:0d", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa150e5dc-12", "ovs_interfaceid": "a150e5dc-1294-45e5-9346-d31c443b7efe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 774.918334] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3bc1e-59d9-3e61-9c65-85a3ec5eff39, 'name': SearchDatastore_Task, 'duration_secs': 0.011842} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.918334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 774.918334] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f521e37d-aa86-4533-9fbe-7d040ed90e53/f521e37d-aa86-4533-9fbe-7d040ed90e53.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 774.918563] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-877027fb-d0dd-498a-bacc-464dc4f64b25 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.928633] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 774.928633] env[65385]: value = "task-4453587" [ 774.928633] env[65385]: _type = "Task" [ 774.928633] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.938804] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453586, 'name': ReconfigVM_Task, 'duration_secs': 0.931387} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.940061] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Reconfigured VM instance instance-00000030 to attach disk [datastore1] b702ba27-bbc7-4030-a96d-30e1ed656a9b/b702ba27-bbc7-4030-a96d-30e1ed656a9b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 774.940823] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad94e44e-de67-4f79-9df0-713c41224b9b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.947581] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.956739] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 774.956739] env[65385]: value = "task-4453588" [ 774.956739] env[65385]: _type = "Task" [ 774.956739] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.970241] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453588, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.008631] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.851s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.037225] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b8da72-afd1-43b6-8923-d6cc6cf6b452 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.051224] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa91f76-ca70-4c65-b873-50c139af9057 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.085463] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27f77d7-1685-42ba-be81-d10fa02b5075 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.094822] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23821d7-34b4-4ecb-909d-344d5c950fc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.114570] env[65385]: DEBUG nova.compute.provider_tree [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.215369] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.259469] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.405156] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "refresh_cache-98430d7e-626e-4532-802e-38f7fc583ae3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 775.405640] env[65385]: DEBUG nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Instance network_info: |[{"id": "a150e5dc-1294-45e5-9346-d31c443b7efe", "address": "fa:16:3e:c8:5a:0d", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa150e5dc-12", "ovs_interfaceid": "a150e5dc-1294-45e5-9346-d31c443b7efe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 775.445551] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453587, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.473521] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453588, 'name': Rename_Task, 'duration_secs': 0.201917} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.473982] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 775.474366] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd968c3a-7c46-4a4a-b64e-3c39c275c997 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.486450] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 775.486450] env[65385]: value = "task-4453589" [ 775.486450] env[65385]: _type = "Task" [ 775.486450] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.499846] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453589, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.512815] env[65385]: DEBUG nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 775.619330] env[65385]: DEBUG nova.scheduler.client.report [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 775.834574] env[65385]: DEBUG nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 775.948211] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609098} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.948564] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f521e37d-aa86-4533-9fbe-7d040ed90e53/f521e37d-aa86-4533-9fbe-7d040ed90e53.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 775.948754] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 775.949141] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dffce512-553c-4a9b-abe6-c32f8b721b60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.958661] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 775.958661] env[65385]: value = "task-4453590" [ 775.958661] env[65385]: _type = "Task" [ 775.958661] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.970271] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453590, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.000442] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453589, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.039690] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.130282] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.321s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.130980] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.423s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.133418] env[65385]: INFO nova.compute.claims [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.161314] env[65385]: INFO nova.scheduler.client.report [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleted allocations for instance 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64 [ 776.472235] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453590, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090511} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.472595] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.474310] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8a201f-1e77-4be2-bef6-243b8dd5ce0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.502155] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] f521e37d-aa86-4533-9fbe-7d040ed90e53/f521e37d-aa86-4533-9fbe-7d040ed90e53.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.503173] env[65385]: DEBUG nova.network.neutron [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Successfully updated port: 3617c8a9-2863-4483-86be-36b932f830a3 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 776.508278] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-761b4d4b-2be3-47f0-b6f3-57de9cd5401e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.532473] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453589, 'name': PowerOnVM_Task, 'duration_secs': 0.724321} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.534187] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 776.534482] env[65385]: INFO nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Took 9.81 seconds to spawn the instance on the hypervisor. [ 776.534669] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 776.535023] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 776.535023] env[65385]: value = "task-4453591" [ 776.535023] env[65385]: _type = "Task" [ 776.535023] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.535774] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c43350e-1817-4fb6-a287-e8c0ce30642b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.557884] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453591, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.673323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36540ed6-c397-4772-a496-c462af1badcd tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.081s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.025622] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "refresh_cache-c5d00bb5-d869-466f-826e-61622e57959f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.026257] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "refresh_cache-c5d00bb5-d869-466f-826e-61622e57959f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.026257] env[65385]: DEBUG nova.network.neutron [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 777.051897] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453591, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.060729] env[65385]: INFO nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Took 44.36 seconds to build instance. [ 777.529793] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.530417] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.556124] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453591, 'name': ReconfigVM_Task, 'duration_secs': 0.605161} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.556488] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Reconfigured VM instance instance-00000031 to attach disk [datastore1] f521e37d-aa86-4533-9fbe-7d040ed90e53/f521e37d-aa86-4533-9fbe-7d040ed90e53.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.557306] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbfdf062-55e4-41f6-942a-1603b02494e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.565890] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.373s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.574306] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 777.574306] env[65385]: value = "task-4453592" [ 777.574306] env[65385]: _type = "Task" [ 777.574306] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.592027] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453592, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.597377] env[65385]: DEBUG nova.network.neutron [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 777.624963] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.625529] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.710558] env[65385]: WARNING neutronclient.v2_0.client [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 777.710558] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.710908] env[65385]: WARNING openstack [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.806957] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09cb825-df58-45ad-a16a-74978828aa0f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.821285] env[65385]: DEBUG nova.network.neutron [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Updating instance_info_cache with network_info: [{"id": "3617c8a9-2863-4483-86be-36b932f830a3", "address": "fa:16:3e:02:12:e8", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3617c8a9-28", "ovs_interfaceid": "3617c8a9-2863-4483-86be-36b932f830a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.824466] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9752cb60-bdfb-489b-9fa6-2262c57bb60e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.859312] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcdb9d1-5d07-4b3b-a378-1f900053fd2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.869908] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0434c2d9-aa6b-4332-a75b-20b6f71a9c64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.888136] env[65385]: DEBUG nova.compute.provider_tree [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.018838] env[65385]: DEBUG nova.compute.manager [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Received event network-changed-a150e5dc-1294-45e5-9346-d31c443b7efe {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 778.018838] env[65385]: DEBUG nova.compute.manager [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Refreshing instance network info cache due to event network-changed-a150e5dc-1294-45e5-9346-d31c443b7efe. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 778.018838] env[65385]: DEBUG oslo_concurrency.lockutils [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Acquiring lock "refresh_cache-98430d7e-626e-4532-802e-38f7fc583ae3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.018838] env[65385]: DEBUG oslo_concurrency.lockutils [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Acquired lock "refresh_cache-98430d7e-626e-4532-802e-38f7fc583ae3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.020918] env[65385]: DEBUG nova.network.neutron [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Refreshing network info cache for port a150e5dc-1294-45e5-9346-d31c443b7efe {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 778.064816] env[65385]: DEBUG nova.compute.manager [req-4806dc29-1390-404e-aac8-476632125b13 req-b14da9af-4c4c-41f7-b9f9-24516063e397 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Received event network-vif-plugged-3617c8a9-2863-4483-86be-36b932f830a3 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 778.065271] env[65385]: DEBUG oslo_concurrency.lockutils [req-4806dc29-1390-404e-aac8-476632125b13 req-b14da9af-4c4c-41f7-b9f9-24516063e397 service nova] Acquiring lock "c5d00bb5-d869-466f-826e-61622e57959f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.065271] env[65385]: DEBUG oslo_concurrency.lockutils [req-4806dc29-1390-404e-aac8-476632125b13 req-b14da9af-4c4c-41f7-b9f9-24516063e397 service nova] Lock "c5d00bb5-d869-466f-826e-61622e57959f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.065579] env[65385]: DEBUG oslo_concurrency.lockutils [req-4806dc29-1390-404e-aac8-476632125b13 req-b14da9af-4c4c-41f7-b9f9-24516063e397 service nova] Lock "c5d00bb5-d869-466f-826e-61622e57959f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.065579] env[65385]: DEBUG nova.compute.manager [req-4806dc29-1390-404e-aac8-476632125b13 req-b14da9af-4c4c-41f7-b9f9-24516063e397 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] No waiting events found dispatching network-vif-plugged-3617c8a9-2863-4483-86be-36b932f830a3 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 778.065714] env[65385]: WARNING nova.compute.manager [req-4806dc29-1390-404e-aac8-476632125b13 req-b14da9af-4c4c-41f7-b9f9-24516063e397 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Received unexpected event network-vif-plugged-3617c8a9-2863-4483-86be-36b932f830a3 for instance with vm_state building and task_state spawning. [ 778.075938] env[65385]: DEBUG nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 778.090185] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453592, 'name': Rename_Task, 'duration_secs': 0.219581} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.091085] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 778.091085] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da65d422-afb1-4959-85b7-383652f6b9b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.100371] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 778.100371] env[65385]: value = "task-4453593" [ 778.100371] env[65385]: _type = "Task" [ 778.100371] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.111103] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.324447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "refresh_cache-c5d00bb5-d869-466f-826e-61622e57959f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 778.325325] env[65385]: DEBUG nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Instance network_info: |[{"id": "3617c8a9-2863-4483-86be-36b932f830a3", "address": "fa:16:3e:02:12:e8", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3617c8a9-28", "ovs_interfaceid": "3617c8a9-2863-4483-86be-36b932f830a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 778.391683] env[65385]: DEBUG nova.scheduler.client.report [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 778.524134] env[65385]: WARNING neutronclient.v2_0.client [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.525087] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.525583] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.614055] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453593, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.622791] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.642507] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.643011] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.731748] env[65385]: WARNING neutronclient.v2_0.client [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.732443] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.732800] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.837022] env[65385]: DEBUG nova.network.neutron [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Updated VIF entry in instance network info cache for port a150e5dc-1294-45e5-9346-d31c443b7efe. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 778.838283] env[65385]: DEBUG nova.network.neutron [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Updating instance_info_cache with network_info: [{"id": "a150e5dc-1294-45e5-9346-d31c443b7efe", "address": "fa:16:3e:c8:5a:0d", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa150e5dc-12", "ovs_interfaceid": "a150e5dc-1294-45e5-9346-d31c443b7efe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 778.899121] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.769s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.900481] env[65385]: DEBUG nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 778.903515] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 31.242s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.114772] env[65385]: DEBUG oslo_vmware.api [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453593, 'name': PowerOnVM_Task, 'duration_secs': 0.743531} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.115262] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 779.115931] env[65385]: INFO nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Took 9.24 seconds to spawn the instance on the hypervisor. [ 779.116348] env[65385]: DEBUG nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 779.117462] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f0e617-6c83-4c9f-9f0a-6f016d3a45ae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.341082] env[65385]: DEBUG oslo_concurrency.lockutils [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Releasing lock "refresh_cache-98430d7e-626e-4532-802e-38f7fc583ae3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.341475] env[65385]: DEBUG nova.compute.manager [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Received event network-changed-94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 779.341811] env[65385]: DEBUG nova.compute.manager [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Refreshing instance network info cache due to event network-changed-94ad8d4c-044d-465c-a0f7-b58c7f70cc14. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 779.342063] env[65385]: DEBUG oslo_concurrency.lockutils [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Acquiring lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.342280] env[65385]: DEBUG oslo_concurrency.lockutils [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Acquired lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.342491] env[65385]: DEBUG nova.network.neutron [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Refreshing network info cache for port 94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 779.409582] env[65385]: INFO nova.compute.claims [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.415159] env[65385]: DEBUG nova.compute.utils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 779.417393] env[65385]: DEBUG nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 779.417564] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 779.417952] env[65385]: WARNING neutronclient.v2_0.client [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 779.418203] env[65385]: WARNING neutronclient.v2_0.client [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 779.418842] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.419436] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.499898] env[65385]: DEBUG nova.policy [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01c9e45d7e6c43eb920276c99d5a2e9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21132a4099d84aaa89698915f7e91c01', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 779.639328] env[65385]: INFO nova.compute.manager [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Took 41.76 seconds to build instance. [ 779.845780] env[65385]: WARNING neutronclient.v2_0.client [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 779.846634] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.847130] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.858322] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Successfully created port: 7217d12a-b81f-4236-971b-a59c2709cf6f {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 779.919543] env[65385]: INFO nova.compute.resource_tracker [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating resource usage from migration 40af7784-05e8-4465-ba0d-2e13af06eb72 [ 779.928404] env[65385]: DEBUG nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 779.992100] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.992509] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.055461] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "649b4e21-fc55-415c-a1f3-ec724397b874" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.055461] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.055776] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "649b4e21-fc55-415c-a1f3-ec724397b874-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.056053] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.056364] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.059624] env[65385]: INFO nova.compute.manager [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Terminating instance [ 780.075393] env[65385]: WARNING neutronclient.v2_0.client [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 780.076169] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.076558] env[65385]: WARNING openstack [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.141253] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8fa4e16b-0333-4666-a7ae-bf0d6a35a743 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.887s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.200280] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Successfully created port: 6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 780.219260] env[65385]: DEBUG nova.network.neutron [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Updated VIF entry in instance network info cache for port 94ad8d4c-044d-465c-a0f7-b58c7f70cc14. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 780.219680] env[65385]: DEBUG nova.network.neutron [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Updating instance_info_cache with network_info: [{"id": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "address": "fa:16:3e:d4:b5:c0", "network": {"id": "08330c64-94a6-41a4-b2a9-1641b6caaf43", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-741669912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a80b4a1e9964a1da68fddeba39f07e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ad8d4c-04", "ovs_interfaceid": "94ad8d4c-044d-465c-a0f7-b58c7f70cc14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.492504] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 780.492743] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 780.492899] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 780.493128] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 780.493293] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 780.493436] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 780.493644] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.493870] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 780.493996] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 780.494248] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 780.494358] env[65385]: DEBUG nova.virt.hardware [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 780.497163] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cb479a-76d5-45a0-b0fb-69ab269909d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.512234] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 780.512590] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 780.512761] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 780.512936] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 780.513095] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 780.513250] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 780.513504] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.513730] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 780.513919] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 780.514176] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 780.514359] env[65385]: DEBUG nova.virt.hardware [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 780.519195] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d950d450-e01d-42ad-8cb6-833f1b1c8f88 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.527367] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4112b07a-4733-4a57-b164-b66ee41adfc1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.530708] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.530992] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.531360] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.531559] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.531734] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.535405] env[65385]: INFO nova.compute.manager [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Terminating instance [ 780.551132] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:0f:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6815237d-f565-474d-a3c0-9c675478eb00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2206d06f-505b-4211-9823-29d900820228', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.558898] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 780.564032] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 780.564294] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 780.564509] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 780.564764] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 780.564898] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 780.565053] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 780.565268] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.565419] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 780.565576] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 780.565730] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 780.565897] env[65385]: DEBUG nova.virt.hardware [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 780.569777] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 780.572073] env[65385]: DEBUG nova.compute.manager [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 780.572153] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.572941] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c092f46-41f4-4d0d-9af3-778772d6d1a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.578098] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7cccb80-b3f8-446d-a370-6718777a43d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.593531] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef155ed6-7f67-4450-987d-59135fa215f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.599258] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3246f8f7-00d1-43ff-9449-1917783f6957 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.610882] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c362a-92de-47ee-9954-186aef65e23e/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 780.611631] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed1dabf-943c-4972-b188-2c5990fce6f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.624548] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461e4d13-5a77-4776-bc6d-7b06cd3e39ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.641626] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:5a:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b89fd3b-0470-40c9-bb5b-d52c76c030e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a150e5dc-1294-45e5-9346-d31c443b7efe', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.649316] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 780.649961] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.649961] env[65385]: value = "task-4453594" [ 780.649961] env[65385]: _type = "Task" [ 780.649961] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.650259] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.653766] env[65385]: DEBUG nova.compute.manager [None req-c1cdca9c-74ad-4fb2-b66b-d923bd58253d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 7bdc813e-5e89-43e2-8963-502f26c998e7] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 780.656579] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 780.656922] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-801f038a-0712-4690-8189-9c08ad887f0f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.670724] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ad86cb5-4d88-484a-971d-cd82a656e7bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.690568] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:12:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3617c8a9-2863-4483-86be-36b932f830a3', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.700044] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 780.700044] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c362a-92de-47ee-9954-186aef65e23e/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 780.700044] env[65385]: ERROR oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c362a-92de-47ee-9954-186aef65e23e/disk-0.vmdk due to incomplete transfer. [ 780.705210] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 780.705484] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-db1cb578-fbac-4672-b5e2-4173a0c21a1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.709277] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc5e411c-8e04-4833-abda-4c0d1e9be98e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.728428] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a7a1c1-d07d-5f5b-ca3d-22a8016137e1/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 780.735460] env[65385]: DEBUG oslo_concurrency.lockutils [req-20dc9d8f-b548-4be7-850c-35555ad5aa42 req-235a973f-8639-45a3-8a40-58916b3a0e5a service nova] Releasing lock "refresh_cache-1ad1e02b-fb82-4a88-9047-be316d590bd2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 780.736353] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.736353] env[65385]: value = "task-4453596" [ 780.736353] env[65385]: _type = "Task" [ 780.736353] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.737510] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291362d4-9786-419e-838b-3f49bc332b57 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.740451] env[65385]: DEBUG oslo_vmware.api [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 780.740451] env[65385]: value = "task-4453595" [ 780.740451] env[65385]: _type = "Task" [ 780.740451] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.740938] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453594, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.752437] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a7a1c1-d07d-5f5b-ca3d-22a8016137e1/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 780.752679] env[65385]: ERROR oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a7a1c1-d07d-5f5b-ca3d-22a8016137e1/disk-0.vmdk due to incomplete transfer. [ 780.752926] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.752926] env[65385]: value = "task-4453597" [ 780.752926] env[65385]: _type = "Task" [ 780.752926] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.754622] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ad2a0e60-b483-41d6-812d-db151a349d6d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.756530] env[65385]: DEBUG oslo_vmware.rw_handles [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c362a-92de-47ee-9954-186aef65e23e/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 780.756887] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Uploaded image 3f16b034-e81d-4f27-a119-fc8ec7cb71a2 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 780.759591] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 780.774789] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6e1edeb0-b0fb-4b92-b568-1ffd75d6d4db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.776575] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453596, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.776962] env[65385]: DEBUG oslo_vmware.api [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453595, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.786450] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453597, 'name': CreateVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.789262] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 780.789262] env[65385]: value = "task-4453598" [ 780.789262] env[65385]: _type = "Task" [ 780.789262] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.790659] env[65385]: DEBUG oslo_vmware.rw_handles [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a7a1c1-d07d-5f5b-ca3d-22a8016137e1/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 780.790659] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Uploaded image 6597b3e2-a8ce-4ef5-81d7-597aba2aa3d6 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 780.791599] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 780.793273] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-321e84ff-6b3e-4c26-a3c1-6fc7ca25f7a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.796146] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f3cc5e-9f98-4e13-a763-c7dc8ab3c795 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.812205] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453598, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.815242] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26388fa5-4b36-4df1-a075-aa226fdbbd2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.819085] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 780.819085] env[65385]: value = "task-4453599" [ 780.819085] env[65385]: _type = "Task" [ 780.819085] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.856988] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42576580-60d2-441d-8815-c067012d238d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.863403] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453599, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.869773] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d6bdeb-9223-4b90-99de-21be5027ce64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.886774] env[65385]: DEBUG nova.compute.provider_tree [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.890723] env[65385]: DEBUG nova.compute.manager [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Received event network-changed-3617c8a9-2863-4483-86be-36b932f830a3 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 780.891000] env[65385]: DEBUG nova.compute.manager [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Refreshing instance network info cache due to event network-changed-3617c8a9-2863-4483-86be-36b932f830a3. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 780.891176] env[65385]: DEBUG oslo_concurrency.lockutils [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Acquiring lock "refresh_cache-c5d00bb5-d869-466f-826e-61622e57959f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.891379] env[65385]: DEBUG oslo_concurrency.lockutils [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Acquired lock "refresh_cache-c5d00bb5-d869-466f-826e-61622e57959f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 780.891548] env[65385]: DEBUG nova.network.neutron [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Refreshing network info cache for port 3617c8a9-2863-4483-86be-36b932f830a3 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 780.947556] env[65385]: DEBUG nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 780.989275] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 780.989623] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 780.989801] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 780.990202] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 780.990202] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 780.990373] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 780.990514] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.990689] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 780.990885] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 780.991051] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 780.991393] env[65385]: DEBUG nova.virt.hardware [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 780.992177] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c42a45-8a3b-4a61-859f-cf3964b199bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.002120] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ac2c7d-0de6-4eee-8773-1b8bc6ca97a5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.071384] env[65385]: DEBUG nova.objects.instance [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lazy-loading 'flavor' on Instance uuid 2f66a03b-874b-4bee-9694-49813ecb0c8a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 781.079453] env[65385]: DEBUG nova.compute.manager [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 781.080887] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.080887] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c5a820-f98d-405d-b8b4-efe6fb553a13 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.092034] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.092182] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-018b8161-e1df-466b-af87-c7c2b732c244 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.169606] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453594, 'name': CreateVM_Task, 'duration_secs': 0.485216} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.170234] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 781.170365] env[65385]: WARNING neutronclient.v2_0.client [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 781.170739] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.170940] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.171337] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 781.171834] env[65385]: DEBUG nova.compute.manager [None req-c1cdca9c-74ad-4fb2-b66b-d923bd58253d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 7bdc813e-5e89-43e2-8963-502f26c998e7] Instance disappeared before build. {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2483}} [ 781.172914] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fb3c113-9e69-4f44-bbbc-8f1df5743868 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.180877] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 781.180877] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52df13fe-8515-499f-835d-09ea286bbc9b" [ 781.180877] env[65385]: _type = "Task" [ 781.180877] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.186099] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.186299] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.186479] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Deleting the datastore file [datastore2] 4232f3a4-c42b-4b9c-8e1d-da591634ffbb {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.187230] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a6baf9d-d6ed-4eda-bba4-93068fb4da16 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.193282] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52df13fe-8515-499f-835d-09ea286bbc9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.197089] env[65385]: DEBUG oslo_vmware.api [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 781.197089] env[65385]: value = "task-4453601" [ 781.197089] env[65385]: _type = "Task" [ 781.197089] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.208985] env[65385]: DEBUG oslo_vmware.api [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453601, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.254831] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453596, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.258661] env[65385]: DEBUG oslo_vmware.api [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453595, 'name': PowerOffVM_Task, 'duration_secs': 0.341616} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.262419] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.262730] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.263429] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6db1edad-1354-4344-9e64-2aa5efb0319f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.273504] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453597, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.304350] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453598, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.336764] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453599, 'name': Destroy_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.357406] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.357673] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.357839] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleting the datastore file [datastore2] 649b4e21-fc55-415c-a1f3-ec724397b874 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.358250] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a5583bd-a74e-4bcb-a747-db283bb18b1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.367236] env[65385]: DEBUG oslo_vmware.api [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for the task: (returnval){ [ 781.367236] env[65385]: value = "task-4453603" [ 781.367236] env[65385]: _type = "Task" [ 781.367236] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.377820] env[65385]: DEBUG oslo_vmware.api [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.390480] env[65385]: DEBUG nova.scheduler.client.report [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 781.395437] env[65385]: WARNING neutronclient.v2_0.client [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.396316] env[65385]: WARNING openstack [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.396444] env[65385]: WARNING openstack [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.577168] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.577381] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquired lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.577718] env[65385]: WARNING neutronclient.v2_0.client [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.578325] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.578671] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.688046] env[65385]: WARNING openstack [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.688506] env[65385]: WARNING openstack [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.695997] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c1cdca9c-74ad-4fb2-b66b-d923bd58253d tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "7bdc813e-5e89-43e2-8963-502f26c998e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.624s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.708229] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52df13fe-8515-499f-835d-09ea286bbc9b, 'name': SearchDatastore_Task, 'duration_secs': 0.014286} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.709014] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.709331] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.709646] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.709781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.709984] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.710301] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ddcb3c8-a84c-448e-b4b6-394601f7c57b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.716583] env[65385]: DEBUG oslo_vmware.api [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453601, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22055} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.717409] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.717616] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.717817] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.718036] env[65385]: INFO nova.compute.manager [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Took 0.64 seconds to destroy the instance on the hypervisor. [ 781.718288] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 781.718790] env[65385]: DEBUG nova.compute.manager [-] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 781.718890] env[65385]: DEBUG nova.network.neutron [-] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 781.719200] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.719862] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.720169] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.728913] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.729124] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.729880] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4869bc05-1327-45d1-be88-96873715291d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.737608] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 781.737608] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f0b8fa-32f5-b12d-d47f-4d846cae8e5e" [ 781.737608] env[65385]: _type = "Task" [ 781.737608] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.749548] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f0b8fa-32f5-b12d-d47f-4d846cae8e5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.756718] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453596, 'name': CreateVM_Task, 'duration_secs': 0.614144} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.756925] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 781.757551] env[65385]: WARNING neutronclient.v2_0.client [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.757994] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.758179] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.758514] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 781.759259] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2027235a-59c2-4202-9b34-c073f6d4ba08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.768983] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 781.768983] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa2891-c005-f333-2915-21385443bc26" [ 781.768983] env[65385]: _type = "Task" [ 781.768983] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.772713] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453597, 'name': CreateVM_Task, 'duration_secs': 0.622706} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.776133] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 781.779415] env[65385]: WARNING neutronclient.v2_0.client [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.780074] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.786765] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa2891-c005-f333-2915-21385443bc26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.805941] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453598, 'name': Destroy_Task, 'duration_secs': 0.652838} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.806382] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Destroyed the VM [ 781.806634] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 781.806948] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-94cde242-e2a4-4782-ae6c-345ac5dc1fd8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.816230] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 781.816230] env[65385]: value = "task-4453604" [ 781.816230] env[65385]: _type = "Task" [ 781.816230] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.838422] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453599, 'name': Destroy_Task, 'duration_secs': 0.53555} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.838671] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453604, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.838908] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Destroyed the VM [ 781.839171] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 781.839432] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9c47f138-f90f-420b-97f4-fa8baeb49e56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.846855] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 781.846855] env[65385]: value = "task-4453605" [ 781.846855] env[65385]: _type = "Task" [ 781.846855] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.858077] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453605, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.881634] env[65385]: DEBUG oslo_vmware.api [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Task: {'id': task-4453603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254898} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.881634] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.881634] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.881634] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.881634] env[65385]: INFO nova.compute.manager [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Took 1.31 seconds to destroy the instance on the hypervisor. [ 781.881634] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 781.883111] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Successfully updated port: 7217d12a-b81f-4236-971b-a59c2709cf6f {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 781.885205] env[65385]: DEBUG nova.compute.manager [-] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 781.885205] env[65385]: DEBUG nova.network.neutron [-] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 781.885205] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.886051] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.886051] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.896847] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.905695] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.003s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.905977] env[65385]: INFO nova.compute.manager [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Migrating [ 781.913817] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.028s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.915755] env[65385]: INFO nova.compute.claims [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.960184] env[65385]: WARNING neutronclient.v2_0.client [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.960948] env[65385]: WARNING openstack [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.961364] env[65385]: WARNING openstack [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.993966] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.206021] env[65385]: DEBUG nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 782.249349] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f0b8fa-32f5-b12d-d47f-4d846cae8e5e, 'name': SearchDatastore_Task, 'duration_secs': 0.019585} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.252253] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adf6ab48-4a06-4820-affc-fb28c2cb6df3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.259867] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 782.259867] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae856a-5e91-56f3-5882-d64b1a62ca60" [ 782.259867] env[65385]: _type = "Task" [ 782.259867] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.273438] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae856a-5e91-56f3-5882-d64b1a62ca60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.288849] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa2891-c005-f333-2915-21385443bc26, 'name': SearchDatastore_Task, 'duration_secs': 0.021219} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.289242] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.289504] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.289733] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.290419] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.290749] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 782.291043] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8379291b-fe99-4196-b472-683b46429589 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.298278] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 782.298278] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cae9ec-5f14-10ac-b98b-a0aabbeb72c9" [ 782.298278] env[65385]: _type = "Task" [ 782.298278] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.308735] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cae9ec-5f14-10ac-b98b-a0aabbeb72c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.309919] env[65385]: DEBUG nova.network.neutron [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 782.330961] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453604, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.364582] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453605, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.432322] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.432322] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.432322] env[65385]: DEBUG nova.network.neutron [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 782.520146] env[65385]: DEBUG nova.network.neutron [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Updated VIF entry in instance network info cache for port 3617c8a9-2863-4483-86be-36b932f830a3. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 782.520610] env[65385]: DEBUG nova.network.neutron [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Updating instance_info_cache with network_info: [{"id": "3617c8a9-2863-4483-86be-36b932f830a3", "address": "fa:16:3e:02:12:e8", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3617c8a9-28", "ovs_interfaceid": "3617c8a9-2863-4483-86be-36b932f830a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 782.734460] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 782.755196] env[65385]: DEBUG nova.network.neutron [-] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 782.772544] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae856a-5e91-56f3-5882-d64b1a62ca60, 'name': SearchDatastore_Task, 'duration_secs': 0.050059} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.773471] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.773805] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 782.774249] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.774355] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.774579] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0c8fd3e-6aa9-4261-947a-d9b263a25bee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.777278] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7277d65-8735-445f-af34-5721638718c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.786519] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 782.786519] env[65385]: value = "task-4453606" [ 782.786519] env[65385]: _type = "Task" [ 782.786519] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.790874] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.791067] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 782.793845] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e529b518-b1b6-4b2b-9cad-83f312ba5ef6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.801068] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453606, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.806287] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 782.806287] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522a9741-50d4-d219-a8c5-5751a1035bb9" [ 782.806287] env[65385]: _type = "Task" [ 782.806287] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.816927] env[65385]: DEBUG nova.compute.manager [req-392df7ad-22d0-44c7-820d-34a89dcf0b9b req-8e84f6be-7fc2-4283-bd0d-40d86e847435 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received event network-vif-plugged-7217d12a-b81f-4236-971b-a59c2709cf6f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 782.817278] env[65385]: DEBUG oslo_concurrency.lockutils [req-392df7ad-22d0-44c7-820d-34a89dcf0b9b req-8e84f6be-7fc2-4283-bd0d-40d86e847435 service nova] Acquiring lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 782.817597] env[65385]: DEBUG oslo_concurrency.lockutils [req-392df7ad-22d0-44c7-820d-34a89dcf0b9b req-8e84f6be-7fc2-4283-bd0d-40d86e847435 service nova] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 782.817800] env[65385]: DEBUG oslo_concurrency.lockutils [req-392df7ad-22d0-44c7-820d-34a89dcf0b9b req-8e84f6be-7fc2-4283-bd0d-40d86e847435 service nova] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.818017] env[65385]: DEBUG nova.compute.manager [req-392df7ad-22d0-44c7-820d-34a89dcf0b9b req-8e84f6be-7fc2-4283-bd0d-40d86e847435 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] No waiting events found dispatching network-vif-plugged-7217d12a-b81f-4236-971b-a59c2709cf6f {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 782.818223] env[65385]: WARNING nova.compute.manager [req-392df7ad-22d0-44c7-820d-34a89dcf0b9b req-8e84f6be-7fc2-4283-bd0d-40d86e847435 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received unexpected event network-vif-plugged-7217d12a-b81f-4236-971b-a59c2709cf6f for instance with vm_state building and task_state spawning. [ 782.818820] env[65385]: WARNING neutronclient.v2_0.client [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.819588] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.821108] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.828412] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cae9ec-5f14-10ac-b98b-a0aabbeb72c9, 'name': SearchDatastore_Task, 'duration_secs': 0.030532} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.829896] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.830161] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.830377] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.838270] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522a9741-50d4-d219-a8c5-5751a1035bb9, 'name': SearchDatastore_Task, 'duration_secs': 0.009944} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.839969] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e8b370a-a7b7-4040-b1bd-bd31747de24d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.847978] env[65385]: DEBUG oslo_vmware.api [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453604, 'name': RemoveSnapshot_Task, 'duration_secs': 0.88272} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.852355] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 782.852626] env[65385]: INFO nova.compute.manager [None req-c21cbdef-1aa3-44a0-9f7b-73478d2bf68f tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Took 17.45 seconds to snapshot the instance on the hypervisor. [ 782.856688] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 782.856688] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5282e7eb-a9a0-5fa3-1e14-5e35a8fe1b4d" [ 782.856688] env[65385]: _type = "Task" [ 782.856688] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.864874] env[65385]: DEBUG oslo_vmware.api [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453605, 'name': RemoveSnapshot_Task, 'duration_secs': 0.752628} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.867950] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 782.869958] env[65385]: INFO nova.compute.manager [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Took 16.49 seconds to snapshot the instance on the hypervisor. [ 782.876384] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5282e7eb-a9a0-5fa3-1e14-5e35a8fe1b4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.935189] env[65385]: WARNING neutronclient.v2_0.client [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.935831] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.936242] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.026012] env[65385]: DEBUG oslo_concurrency.lockutils [req-004daec2-75ca-4255-934e-c1e26f1535f2 req-2ce99dbb-6ea2-4b8d-94bd-10108c2ccf62 service nova] Releasing lock "refresh_cache-c5d00bb5-d869-466f-826e-61622e57959f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.135058] env[65385]: DEBUG nova.network.neutron [-] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.210360] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.210779] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.245916] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.246555] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.258078] env[65385]: INFO nova.compute.manager [-] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Took 1.54 seconds to deallocate network for instance. [ 783.313288] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453606, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.375501] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5282e7eb-a9a0-5fa3-1e14-5e35a8fe1b4d, 'name': SearchDatastore_Task, 'duration_secs': 0.030954} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.384088] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.384603] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 98430d7e-626e-4532-802e-38f7fc583ae3/98430d7e-626e-4532-802e-38f7fc583ae3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 783.388139] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.388378] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.388708] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08c660eb-7e8a-4828-b63a-1d73051e3da0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.394232] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21756be2-6ab1-47f6-bfba-e791b3b2cf64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.407324] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 783.407324] env[65385]: value = "task-4453607" [ 783.407324] env[65385]: _type = "Task" [ 783.407324] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.415895] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.415895] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.418076] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e966b501-bd51-4237-94c9-06efbf1c0c03 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.432339] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453607, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.435037] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 783.435037] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52710590-5318-ee64-bdff-cbd0ada45f45" [ 783.435037] env[65385]: _type = "Task" [ 783.435037] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.450780] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52710590-5318-ee64-bdff-cbd0ada45f45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.469987] env[65385]: DEBUG nova.compute.manager [None req-0235ba82-29d1-4388-a407-81ab58a32bf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Found 2 images (rotation: 2) {{(pid=65385) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 783.498743] env[65385]: WARNING neutronclient.v2_0.client [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.499480] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.499832] env[65385]: WARNING openstack [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.543428] env[65385]: WARNING neutronclient.v2_0.client [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.544199] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.544700] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.633201] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1491f3-479d-46c8-ab34-1044f7ce59ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.641863] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5e723b-c2b1-4e43-82b1-22fa484df6f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.645966] env[65385]: INFO nova.compute.manager [-] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Took 1.76 seconds to deallocate network for instance. [ 783.708111] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022aa169-47c8-46ac-8c73-a9a0e7d02490 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.717905] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df4e42c-0160-4708-b48c-9f135441958f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.734509] env[65385]: DEBUG nova.compute.provider_tree [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.739928] env[65385]: DEBUG nova.network.neutron [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.767160] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.772654] env[65385]: DEBUG nova.network.neutron [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance_info_cache with network_info: [{"id": "ad95d76f-e023-47b2-a936-e7ed11779276", "address": "fa:16:3e:b0:c7:d2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95d76f-e0", "ovs_interfaceid": "ad95d76f-e023-47b2-a936-e7ed11779276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.806998] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453606, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.81751} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.807959] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.808264] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.808580] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e9b6b2cd-f97f-4932-be01-85bd74f743fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.821019] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 783.821019] env[65385]: value = "task-4453608" [ 783.821019] env[65385]: _type = "Task" [ 783.821019] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.833858] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453608, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.921815] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453607, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.950057] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52710590-5318-ee64-bdff-cbd0ada45f45, 'name': SearchDatastore_Task, 'duration_secs': 0.061894} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.951138] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02e80896-7686-4ad9-89f6-198bea128699 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.958648] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 783.958648] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5212f516-2418-4644-2be6-980ad9d47fe9" [ 783.958648] env[65385]: _type = "Task" [ 783.958648] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.970905] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5212f516-2418-4644-2be6-980ad9d47fe9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.057256] env[65385]: DEBUG nova.compute.manager [req-e06d3900-85a5-48da-8aa7-4614ced894a0 req-46089b89-7c57-401c-b612-714f670befe8 service nova] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Received event network-vif-deleted-b74e7138-b571-449e-af1d-9e68bf8e9064 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 784.127774] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Successfully updated port: 6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 784.207115] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.242076] env[65385]: DEBUG nova.scheduler.client.report [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 784.245679] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Releasing lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 784.245936] env[65385]: DEBUG nova.compute.manager [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Inject network info {{(pid=65385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7777}} [ 784.246165] env[65385]: DEBUG nova.compute.manager [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] network_info to inject: |[{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7778}} [ 784.251199] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Reconfiguring VM instance to set the machine id {{(pid=65385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 784.251786] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e5f64bb-0a1b-4f91-9ca2-ca0c629301f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.269677] env[65385]: DEBUG oslo_vmware.api [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 784.269677] env[65385]: value = "task-4453609" [ 784.269677] env[65385]: _type = "Task" [ 784.269677] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.280165] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 784.281513] env[65385]: DEBUG oslo_vmware.api [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453609, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.298980] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "e3248f4a-8985-48b0-84cf-795f00ff21fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.299409] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.299662] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "e3248f4a-8985-48b0-84cf-795f00ff21fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.299884] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.300157] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 784.302343] env[65385]: INFO nova.compute.manager [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Terminating instance [ 784.333073] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453608, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098801} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.333363] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.334302] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0080fbc5-b61d-40c6-9116-3c1b9fe3cff8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.363815] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.364509] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebfc56a9-1184-415d-913e-ff901fc35ea0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.388637] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 784.388637] env[65385]: value = "task-4453610" [ 784.388637] env[65385]: _type = "Task" [ 784.388637] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.398898] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.421428] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453607, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653271} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.421428] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 98430d7e-626e-4532-802e-38f7fc583ae3/98430d7e-626e-4532-802e-38f7fc583ae3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 784.421428] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.421428] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ba796b0-6836-4067-ba21-d48c252f606e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.430202] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 784.430202] env[65385]: value = "task-4453611" [ 784.430202] env[65385]: _type = "Task" [ 784.430202] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.440438] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.471875] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5212f516-2418-4644-2be6-980ad9d47fe9, 'name': SearchDatastore_Task, 'duration_secs': 0.051733} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.472226] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 784.472460] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c5d00bb5-d869-466f-826e-61622e57959f/c5d00bb5-d869-466f-826e-61622e57959f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.472741] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8b82e7f-c416-4a60-994b-05cdf164300f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.481789] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 784.481789] env[65385]: value = "task-4453612" [ 784.481789] env[65385]: _type = "Task" [ 784.481789] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.492519] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453612, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.630300] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.630300] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 784.630300] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 784.755456] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.841s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 784.756153] env[65385]: DEBUG nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 784.759140] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.020s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.759320] env[65385]: DEBUG nova.objects.instance [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lazy-loading 'resources' on Instance uuid 2e4c064f-27b0-4ccb-8fe6-c34a61153663 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 784.790664] env[65385]: DEBUG oslo_vmware.api [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453609, 'name': ReconfigVM_Task, 'duration_secs': 0.186987} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.790664] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d83fafdd-c643-421b-8a9f-3ac39380c607 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Reconfigured VM instance to set the machine id {{(pid=65385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 784.806632] env[65385]: DEBUG nova.compute.manager [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 784.807070] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.808116] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc0dd94-781a-4b41-bca1-3f39a2a95d4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.822947] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 784.824013] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34f2f5b5-d138-40e2-97b0-c05b90f7ab54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.833716] env[65385]: DEBUG oslo_vmware.api [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 784.833716] env[65385]: value = "task-4453613" [ 784.833716] env[65385]: _type = "Task" [ 784.833716] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.850036] env[65385]: DEBUG oslo_vmware.api [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453613, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.902186] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453610, 'name': ReconfigVM_Task, 'duration_secs': 0.398898} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.902621] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Reconfigured VM instance instance-00000014 to attach disk [datastore2] eb9d0510-b453-4695-9e1d-731217b9f8ec/eb9d0510-b453-4695-9e1d-731217b9f8ec.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.903426] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d20a56b-2589-497c-aad9-9fe085b42766 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.916623] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 784.916623] env[65385]: value = "task-4453614" [ 784.916623] env[65385]: _type = "Task" [ 784.916623] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.928265] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453614, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.944402] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089289} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.945270] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.945606] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8a53cf-85df-4b3e-85cd-fe56324ee768 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.980851] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 98430d7e-626e-4532-802e-38f7fc583ae3/98430d7e-626e-4532-802e-38f7fc583ae3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.981724] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe33a668-4fd4-4207-b2f5-f570f8bbc7fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.017707] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453612, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.019608] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 785.019608] env[65385]: value = "task-4453615" [ 785.019608] env[65385]: _type = "Task" [ 785.019608] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.032076] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453615, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.133643] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.134162] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.179770] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 785.194378] env[65385]: DEBUG nova.objects.instance [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lazy-loading 'flavor' on Instance uuid 2f66a03b-874b-4bee-9694-49813ecb0c8a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 785.200884] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.201321] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.263175] env[65385]: DEBUG nova.compute.utils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 785.267844] env[65385]: DEBUG nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 785.268142] env[65385]: DEBUG nova.network.neutron [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 785.268478] env[65385]: WARNING neutronclient.v2_0.client [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 785.268789] env[65385]: WARNING neutronclient.v2_0.client [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 785.269414] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.269781] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.280958] env[65385]: WARNING neutronclient.v2_0.client [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 785.281407] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.284761] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.336338] env[65385]: DEBUG nova.policy [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e9fa9ad20645f59a103ec416ee8666', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d75a2e244de46358945f6f2a57f1efc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 785.358878] env[65385]: DEBUG oslo_vmware.api [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453613, 'name': PowerOffVM_Task, 'duration_secs': 0.266661} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.358878] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 785.358878] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 785.359169] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af8686be-ac2a-4e69-8995-eb0a963e192d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.399835] env[65385]: DEBUG nova.compute.manager [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received event network-changed-7217d12a-b81f-4236-971b-a59c2709cf6f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 785.401047] env[65385]: DEBUG nova.compute.manager [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Refreshing instance network info cache due to event network-changed-7217d12a-b81f-4236-971b-a59c2709cf6f. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 785.401047] env[65385]: DEBUG oslo_concurrency.lockutils [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Acquiring lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.435848] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453614, 'name': Rename_Task, 'duration_secs': 0.268955} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.436351] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 785.436628] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eccfe6df-5ae9-480a-b854-368738ef09f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.441115] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 785.441616] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 785.441616] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleting the datastore file [datastore1] e3248f4a-8985-48b0-84cf-795f00ff21fa {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.443101] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.443664] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.450066] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f35e5332-e126-40a9-8174-9fb660c357ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.455757] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 785.455757] env[65385]: value = "task-4453617" [ 785.455757] env[65385]: _type = "Task" [ 785.455757] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.461598] env[65385]: DEBUG oslo_vmware.api [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 785.461598] env[65385]: value = "task-4453618" [ 785.461598] env[65385]: _type = "Task" [ 785.461598] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.469933] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453617, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.476326] env[65385]: DEBUG oslo_vmware.api [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.516987] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453612, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578717} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.522367] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c5d00bb5-d869-466f-826e-61622e57959f/c5d00bb5-d869-466f-826e-61622e57959f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.523421] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.523421] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ad2cf22-a3b3-4c9e-accb-cc752479eb4f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.540630] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453615, 'name': ReconfigVM_Task, 'duration_secs': 0.365381} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.542727] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 98430d7e-626e-4532-802e-38f7fc583ae3/98430d7e-626e-4532-802e-38f7fc583ae3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.543643] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 785.543643] env[65385]: value = "task-4453619" [ 785.543643] env[65385]: _type = "Task" [ 785.543643] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.544743] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb85930f-8901-49a7-9008-28bf4ab5cb4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.561957] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.562679] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 785.562679] env[65385]: value = "task-4453620" [ 785.562679] env[65385]: _type = "Task" [ 785.562679] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.577076] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453620, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.589505] env[65385]: WARNING neutronclient.v2_0.client [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 785.590232] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.592089] env[65385]: WARNING openstack [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.700315] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.700440] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquired lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 785.701525] env[65385]: WARNING neutronclient.v2_0.client [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 785.744579] env[65385]: DEBUG nova.compute.manager [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 785.746036] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a206b010-1fe8-4b24-ab23-74e18763772f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.763608] env[65385]: DEBUG nova.network.neutron [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Successfully created port: 3ed3260d-d4bf-49a0-baa3-bb749653df5e {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 785.769918] env[65385]: DEBUG nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 785.775588] env[65385]: DEBUG nova.network.neutron [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Updating instance_info_cache with network_info: [{"id": "7217d12a-b81f-4236-971b-a59c2709cf6f", "address": "fa:16:3e:c4:68:5c", "network": {"id": "4ff82148-f4e8-42a8-853a-d8d163ae26bd", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1544056117", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7217d12a-b8", "ovs_interfaceid": "7217d12a-b81f-4236-971b-a59c2709cf6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "address": "fa:16:3e:a6:d3:58", "network": {"id": "9159387d-e138-4204-8ed9-ac0e0da90550", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-474388752", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f259efd-d3", "ovs_interfaceid": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 785.804105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7620f6cb-848e-45e1-90e5-4ab6e0560784 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.832598] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance '38e0ebf5-1d92-4906-b0bd-b0ebf8651f22' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 785.961872] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42fdb90-1dd0-41e3-a10c-a81e4f187cb0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.972740] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453617, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.976625] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcebb72-edc0-4a46-a2b7-ad5e8707f985 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.983970] env[65385]: DEBUG oslo_vmware.api [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.984835] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.985170] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 785.985434] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.985674] env[65385]: INFO nova.compute.manager [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Took 1.18 seconds to destroy the instance on the hypervisor. [ 785.985893] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 785.986110] env[65385]: DEBUG nova.compute.manager [-] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 785.986203] env[65385]: DEBUG nova.network.neutron [-] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 785.986651] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 785.987113] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.987402] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.021447] env[65385]: DEBUG nova.compute.manager [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 786.023144] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec770c1-81f9-4780-9874-f5faf3b1110d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.026421] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cffcbc5-4bff-4017-aa15-59796c1a06cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.039979] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915e2b2e-a074-4158-8b90-1df179dde9b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.058231] env[65385]: DEBUG nova.compute.provider_tree [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.066734] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077028} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.069283] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.071293] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8efb36-3a22-478a-975d-86fcffae8c7b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.083416] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453620, 'name': Rename_Task, 'duration_secs': 0.227579} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.093494] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 786.103519] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] c5d00bb5-d869-466f-826e-61622e57959f/c5d00bb5-d869-466f-826e-61622e57959f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.103867] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-718602b7-ec8a-4838-bae4-3b7df14b8fad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.106282] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-750d477f-4ae3-4330-980c-c955b87e81b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.128136] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 786.128136] env[65385]: value = "task-4453621" [ 786.128136] env[65385]: _type = "Task" [ 786.128136] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.130114] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 786.130114] env[65385]: value = "task-4453622" [ 786.130114] env[65385]: _type = "Task" [ 786.130114] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.142177] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453621, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.145594] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453622, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.152291] env[65385]: DEBUG nova.network.neutron [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 786.230025] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.263334] env[65385]: INFO nova.compute.manager [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] instance snapshotting [ 786.264067] env[65385]: DEBUG nova.objects.instance [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'flavor' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 786.283712] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Releasing lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.283712] env[65385]: DEBUG nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Instance network_info: |[{"id": "7217d12a-b81f-4236-971b-a59c2709cf6f", "address": "fa:16:3e:c4:68:5c", "network": {"id": "4ff82148-f4e8-42a8-853a-d8d163ae26bd", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1544056117", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7217d12a-b8", "ovs_interfaceid": "7217d12a-b81f-4236-971b-a59c2709cf6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "address": "fa:16:3e:a6:d3:58", "network": {"id": "9159387d-e138-4204-8ed9-ac0e0da90550", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-474388752", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f259efd-d3", "ovs_interfaceid": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 786.283952] env[65385]: DEBUG oslo_concurrency.lockutils [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Acquired lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.284062] env[65385]: DEBUG nova.network.neutron [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Refreshing network info cache for port 7217d12a-b81f-4236-971b-a59c2709cf6f {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 786.288951] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:68:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b0fa7a2-ebd9-4788-8904-7bf250ce466c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7217d12a-b81f-4236-971b-a59c2709cf6f', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:d3:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dabbac20-1723-40ad-9da0-e53b28073651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f259efd-d33d-4c33-8f8d-a98afc1ec5b8', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 786.295236] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 786.296613] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 786.296613] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1c7bdb3-294f-4155-badb-ac95b2d5b429 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.322717] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 786.322717] env[65385]: value = "task-4453623" [ 786.322717] env[65385]: _type = "Task" [ 786.322717] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.333659] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453623, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.344720] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.345424] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c86845f1-4ec1-4786-bb25-8bbc9d1f76b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.354127] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 786.354127] env[65385]: value = "task-4453624" [ 786.354127] env[65385]: _type = "Task" [ 786.354127] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.365949] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.467947] env[65385]: DEBUG oslo_vmware.api [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453617, 'name': PowerOnVM_Task, 'duration_secs': 0.567757} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.468293] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 786.468494] env[65385]: DEBUG nova.compute.manager [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 786.469371] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed734fa4-1a3f-43f7-8f0c-ec9b6455c075 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.540291] env[65385]: INFO nova.compute.manager [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] instance snapshotting [ 786.544924] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634d0402-0c76-4099-807e-5804606452e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.570747] env[65385]: DEBUG nova.scheduler.client.report [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 786.576402] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c2b611-c6d9-4f9e-bbbd-17fa4847d7bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.643853] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453621, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.647383] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453622, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.656660] env[65385]: WARNING neutronclient.v2_0.client [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.656660] env[65385]: WARNING openstack [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.656660] env[65385]: WARNING openstack [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.771931] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de53276-6c2c-40ed-ada7-f4bdda67630c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.781365] env[65385]: DEBUG nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 786.799976] env[65385]: WARNING neutronclient.v2_0.client [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.800667] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.801037] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.814891] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd484f2-9bf2-4cb6-bd3f-ed569bcba98a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.844873] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 786.845396] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 786.845396] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 786.845574] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 786.845723] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 786.845947] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 786.846237] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.847728] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 786.847728] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 786.847728] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 786.847728] env[65385]: DEBUG nova.virt.hardware [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 786.847728] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453623, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.849555] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659ca424-8e78-4355-a955-4930b5b2d14d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.856160] env[65385]: DEBUG nova.compute.manager [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received event network-vif-plugged-6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 786.856424] env[65385]: DEBUG oslo_concurrency.lockutils [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Acquiring lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.856651] env[65385]: DEBUG oslo_concurrency.lockutils [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.857562] env[65385]: DEBUG oslo_concurrency.lockutils [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.857914] env[65385]: DEBUG nova.compute.manager [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] No waiting events found dispatching network-vif-plugged-6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 786.860685] env[65385]: WARNING nova.compute.manager [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received unexpected event network-vif-plugged-6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 for instance with vm_state building and task_state spawning. [ 786.860685] env[65385]: DEBUG nova.compute.manager [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received event network-changed-6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 786.860685] env[65385]: DEBUG nova.compute.manager [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Refreshing instance network info cache due to event network-changed-6f259efd-d33d-4c33-8f8d-a98afc1ec5b8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 786.860685] env[65385]: DEBUG oslo_concurrency.lockutils [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Acquiring lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.870593] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c502a5-3e30-49ad-9a05-6bc6d193899d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.881752] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453624, 'name': PowerOffVM_Task, 'duration_secs': 0.291246} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.883484] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.883800] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance '38e0ebf5-1d92-4906-b0bd-b0ebf8651f22' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 786.913596] env[65385]: WARNING openstack [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.914068] env[65385]: WARNING openstack [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.988477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.011633] env[65385]: DEBUG nova.network.neutron [-] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.055451] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.055838] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.081307] env[65385]: WARNING neutronclient.v2_0.client [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.081969] env[65385]: WARNING openstack [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.082326] env[65385]: WARNING openstack [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.090296] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.331s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.092724] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 787.093362] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.033s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.093574] env[65385]: DEBUG nova.objects.instance [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lazy-loading 'resources' on Instance uuid fae095a8-a768-4263-b456-ed0745398755 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.094911] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-66cbccb5-c054-4f6b-a313-167c5d588ce5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.104219] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 787.104219] env[65385]: value = "task-4453625" [ 787.104219] env[65385]: _type = "Task" [ 787.104219] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.122025] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453625, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.126841] env[65385]: INFO nova.scheduler.client.report [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Deleted allocations for instance 2e4c064f-27b0-4ccb-8fe6-c34a61153663 [ 787.143068] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453621, 'name': ReconfigVM_Task, 'duration_secs': 0.602255} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.143863] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Reconfigured VM instance instance-00000033 to attach disk [datastore2] c5d00bb5-d869-466f-826e-61622e57959f/c5d00bb5-d869-466f-826e-61622e57959f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.144624] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6355b25d-9133-415f-90a5-f2dccf9910e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.150368] env[65385]: DEBUG oslo_vmware.api [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453622, 'name': PowerOnVM_Task, 'duration_secs': 0.68557} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.151033] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 787.151287] env[65385]: INFO nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Took 14.30 seconds to spawn the instance on the hypervisor. [ 787.151527] env[65385]: DEBUG nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 787.152601] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ed29d9-4045-4c0d-818a-2b6cca2afe3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.159024] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 787.159024] env[65385]: value = "task-4453626" [ 787.159024] env[65385]: _type = "Task" [ 787.159024] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.175109] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453626, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.203623] env[65385]: WARNING neutronclient.v2_0.client [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.204340] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.204700] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.252872] env[65385]: DEBUG nova.network.neutron [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.339935] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 787.340410] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453623, 'name': CreateVM_Task, 'duration_secs': 0.627482} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.340665] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-201500e6-a59b-4f2b-9afd-0ad48a8ba781 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.343584] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 787.344773] env[65385]: WARNING neutronclient.v2_0.client [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.345211] env[65385]: WARNING neutronclient.v2_0.client [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.345543] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.345691] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 787.345996] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 787.346662] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87d88148-f405-4ce2-b7b3-859055abedb2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.351976] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 787.351976] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52440da7-40af-a9ea-bd91-5006caa7b006" [ 787.351976] env[65385]: _type = "Task" [ 787.351976] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.353444] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 787.353444] env[65385]: value = "task-4453627" [ 787.353444] env[65385]: _type = "Task" [ 787.353444] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.358838] env[65385]: DEBUG nova.network.neutron [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Updated VIF entry in instance network info cache for port 7217d12a-b81f-4236-971b-a59c2709cf6f. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 787.359286] env[65385]: DEBUG nova.network.neutron [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Updating instance_info_cache with network_info: [{"id": "7217d12a-b81f-4236-971b-a59c2709cf6f", "address": "fa:16:3e:c4:68:5c", "network": {"id": "4ff82148-f4e8-42a8-853a-d8d163ae26bd", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1544056117", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7217d12a-b8", "ovs_interfaceid": "7217d12a-b81f-4236-971b-a59c2709cf6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "address": "fa:16:3e:a6:d3:58", "network": {"id": "9159387d-e138-4204-8ed9-ac0e0da90550", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-474388752", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f259efd-d3", "ovs_interfaceid": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.372018] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52440da7-40af-a9ea-bd91-5006caa7b006, 'name': SearchDatastore_Task, 'duration_secs': 0.015588} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.376162] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.376502] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.376876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.377053] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 787.377278] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.377645] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453627, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.378028] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e32d9e0-1c48-4e8c-8269-1d0220d001b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.390814] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.391066] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 787.391934] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-598d3766-d7c9-446f-acc1-106212a22f75 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.400517] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 787.400863] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 787.401230] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 787.401500] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 787.401697] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 787.401862] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 787.402224] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.402419] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 787.402674] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 787.402945] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 787.403235] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 787.412480] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48bf0fe7-0361-41e5-9c62-16c23d70e9bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.424495] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 787.424495] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52211d23-5fec-7bf1-7648-3886fb3af585" [ 787.424495] env[65385]: _type = "Task" [ 787.424495] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.432801] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 787.432801] env[65385]: value = "task-4453628" [ 787.432801] env[65385]: _type = "Task" [ 787.432801] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.437197] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52211d23-5fec-7bf1-7648-3886fb3af585, 'name': SearchDatastore_Task, 'duration_secs': 0.014314} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.441713] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1260e757-3ed8-4e27-bde4-e0795b9b8824 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.451497] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453628, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.453564] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 787.453564] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528d526a-27ff-0222-f31e-93d6b79aa26a" [ 787.453564] env[65385]: _type = "Task" [ 787.453564] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.465342] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528d526a-27ff-0222-f31e-93d6b79aa26a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.478993] env[65385]: DEBUG nova.network.neutron [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Successfully updated port: 3ed3260d-d4bf-49a0-baa3-bb749653df5e {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 787.514929] env[65385]: INFO nova.compute.manager [-] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Took 1.53 seconds to deallocate network for instance. [ 787.617191] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453625, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.638712] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dfeb5b61-fb98-427f-8096-d4877ebcbb8b tempest-ServersV294TestFqdnHostnames-1069356813 tempest-ServersV294TestFqdnHostnames-1069356813-project-member] Lock "2e4c064f-27b0-4ccb-8fe6-c34a61153663" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.783s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.677024] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453626, 'name': Rename_Task, 'duration_secs': 0.170247} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.679087] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.679635] env[65385]: INFO nova.compute.manager [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Took 49.72 seconds to build instance. [ 787.680711] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ff9f1e7-75b9-40f9-9872-de48af88181a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.693472] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 787.693472] env[65385]: value = "task-4453629" [ 787.693472] env[65385]: _type = "Task" [ 787.693472] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.705937] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.756109] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Releasing lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.756363] env[65385]: DEBUG nova.compute.manager [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Inject network info {{(pid=65385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7777}} [ 787.756718] env[65385]: DEBUG nova.compute.manager [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] network_info to inject: |[{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7778}} [ 787.763020] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Reconfiguring VM instance to set the machine id {{(pid=65385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 787.763020] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-227de6c5-63c3-49ac-86b0-38437a724708 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.789290] env[65385]: DEBUG oslo_vmware.api [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 787.789290] env[65385]: value = "task-4453630" [ 787.789290] env[65385]: _type = "Task" [ 787.789290] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.804756] env[65385]: DEBUG oslo_vmware.api [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453630, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.826436] env[65385]: DEBUG nova.compute.manager [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Received event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 787.826786] env[65385]: DEBUG nova.compute.manager [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing instance network info cache due to event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 787.827396] env[65385]: DEBUG oslo_concurrency.lockutils [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Acquiring lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.827688] env[65385]: DEBUG oslo_concurrency.lockutils [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Acquired lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 787.827949] env[65385]: DEBUG nova.network.neutron [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 787.865876] env[65385]: DEBUG oslo_concurrency.lockutils [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Releasing lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.866081] env[65385]: DEBUG nova.compute.manager [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Received event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 787.866535] env[65385]: DEBUG nova.compute.manager [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing instance network info cache due to event network-changed-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 787.866535] env[65385]: DEBUG oslo_concurrency.lockutils [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Acquiring lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.872120] env[65385]: DEBUG oslo_concurrency.lockutils [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Acquired lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 787.872120] env[65385]: DEBUG nova.network.neutron [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Refreshing network info cache for port 6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 787.874262] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453627, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.955142] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453628, 'name': ReconfigVM_Task, 'duration_secs': 0.223154} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.960951] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance '38e0ebf5-1d92-4906-b0bd-b0ebf8651f22' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 787.977417] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528d526a-27ff-0222-f31e-93d6b79aa26a, 'name': SearchDatastore_Task, 'duration_secs': 0.014949} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.977720] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.977997] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e9fd0c9f-5795-44df-9dcb-982b7722ae69/e9fd0c9f-5795-44df-9dcb-982b7722ae69.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 787.978283] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3310d30e-d82e-4b07-ac4f-3ff2fe113516 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.982380] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "refresh_cache-ab8df1d8-615d-4083-bdf4-25f679a64982" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.983065] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquired lock "refresh_cache-ab8df1d8-615d-4083-bdf4-25f679a64982" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 787.983065] env[65385]: DEBUG nova.network.neutron [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 787.996496] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 787.996496] env[65385]: value = "task-4453631" [ 787.996496] env[65385]: _type = "Task" [ 787.996496] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.008533] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453631, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.022341] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.120462] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453625, 'name': CreateSnapshot_Task, 'duration_secs': 0.597899} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.123686] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 788.124949] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0696e1-d1d8-42ae-ada9-7ff38f06b5e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.185818] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8a80c108-47c9-4cc8-87b0-1e2f5388a7b2 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "98430d7e-626e-4532-802e-38f7fc583ae3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.777s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.210094] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453629, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.294325] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.294610] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.294889] env[65385]: INFO nova.compute.manager [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Rebooting instance [ 788.299282] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ead52ec-d27e-49ac-97f5-e81b12c0bbb0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.315254] env[65385]: DEBUG oslo_vmware.api [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453630, 'name': ReconfigVM_Task, 'duration_secs': 0.204752} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.317654] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc57bb1-3db4-4416-ac38-b131a4270e1b tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Reconfigured VM instance to set the machine id {{(pid=65385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 788.322907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a3ff72-65dd-48b5-8c77-4a2748e274f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.362047] env[65385]: WARNING neutronclient.v2_0.client [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.362446] env[65385]: WARNING openstack [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.362888] env[65385]: WARNING openstack [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.379883] env[65385]: WARNING neutronclient.v2_0.client [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.381127] env[65385]: WARNING openstack [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.381711] env[65385]: WARNING openstack [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.397569] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93aaf7d-fb7f-40a3-b17e-1ba85b57af52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.412279] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453627, 'name': CreateSnapshot_Task, 'duration_secs': 0.878864} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.413154] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 788.416567] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b5fb2f-be93-451a-b0bb-400c5be764ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.420381] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c263dd4f-4a57-46c5-aa26-32c3d526ee12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.441590] env[65385]: DEBUG nova.compute.provider_tree [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.474356] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:48:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f35b2975-b7d3-4f0c-9c41-d7d5c82d8413',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1230048441',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 788.474596] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 788.474733] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 788.475040] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 788.475262] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 788.475427] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 788.475704] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.475869] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 788.476056] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 788.476222] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 788.476414] env[65385]: DEBUG nova.virt.hardware [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 788.484300] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfiguring VM instance instance-00000029 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 788.488564] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-884054f9-af6e-4313-a6a6-752171372386 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.506488] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.507014] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.514573] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "2f66a03b-874b-4bee-9694-49813ecb0c8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.514829] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.515096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "2f66a03b-874b-4bee-9694-49813ecb0c8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.515290] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.515456] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.523441] env[65385]: INFO nova.compute.manager [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Terminating instance [ 788.531587] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 788.531587] env[65385]: value = "task-4453632" [ 788.531587] env[65385]: _type = "Task" [ 788.531587] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.535440] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453631, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.547189] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453632, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.648479] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 788.648833] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5e38228d-7592-467d-b9e6-43ae98b2bc52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.658404] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 788.658404] env[65385]: value = "task-4453633" [ 788.658404] env[65385]: _type = "Task" [ 788.658404] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.668061] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453633, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.688391] env[65385]: DEBUG nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 788.707497] env[65385]: DEBUG oslo_vmware.api [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453629, 'name': PowerOnVM_Task, 'duration_secs': 0.879025} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.707849] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.708136] env[65385]: INFO nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Took 12.87 seconds to spawn the instance on the hypervisor. [ 788.708348] env[65385]: DEBUG nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 788.709348] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b846b1cc-8618-46ce-8195-cedb50c89c9f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.730600] env[65385]: DEBUG nova.network.neutron [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 788.773435] env[65385]: WARNING openstack [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.773435] env[65385]: WARNING openstack [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.822920] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.826600] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.826819] env[65385]: DEBUG nova.network.neutron [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 788.834716] env[65385]: WARNING openstack [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.835177] env[65385]: WARNING openstack [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.880236] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.880236] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.947252] env[65385]: DEBUG nova.scheduler.client.report [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 788.959156] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 788.959907] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3299e153-f6e1-42fb-889c-85632776d8b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.974891] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 788.974891] env[65385]: value = "task-4453634" [ 788.974891] env[65385]: _type = "Task" [ 788.974891] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.986666] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453634, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.031604] env[65385]: DEBUG nova.compute.manager [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 789.031604] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.031604] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453631, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653753} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.031604] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1a1f01-28b2-4cca-adcf-8b1206fbdfa0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.034989] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e9fd0c9f-5795-44df-9dcb-982b7722ae69/e9fd0c9f-5795-44df-9dcb-982b7722ae69.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 789.035565] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.038594] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be0b737d-03e7-4ca4-889a-5a42b368a796 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.048438] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 789.050508] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f859df00-8b2d-4570-86ba-5c22f4616784 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.055423] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453632, 'name': ReconfigVM_Task, 'duration_secs': 0.288589} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.057220] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfigured VM instance instance-00000029 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 789.057718] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 789.057718] env[65385]: value = "task-4453635" [ 789.057718] env[65385]: _type = "Task" [ 789.057718] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.060032] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f7b59e-500c-48a9-bfcd-b589dc10d01a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.063590] env[65385]: DEBUG oslo_vmware.api [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 789.063590] env[65385]: value = "task-4453636" [ 789.063590] env[65385]: _type = "Task" [ 789.063590] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.074236] env[65385]: WARNING neutronclient.v2_0.client [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.075306] env[65385]: WARNING openstack [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.075924] env[65385]: WARNING openstack [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.106468] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.112168] env[65385]: WARNING neutronclient.v2_0.client [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.113067] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.113593] env[65385]: WARNING openstack [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.126028] env[65385]: WARNING neutronclient.v2_0.client [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.126028] env[65385]: WARNING openstack [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.126028] env[65385]: WARNING openstack [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.133073] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e10066cb-3c35-49c6-9b9e-d849324962f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.149026] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453635, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.153298] env[65385]: DEBUG oslo_vmware.api [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.165214] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 789.165214] env[65385]: value = "task-4453637" [ 789.165214] env[65385]: _type = "Task" [ 789.165214] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.173172] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453633, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.179734] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453637, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.230129] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 789.233588] env[65385]: INFO nova.compute.manager [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Took 44.05 seconds to build instance. [ 789.333273] env[65385]: WARNING neutronclient.v2_0.client [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.334493] env[65385]: WARNING openstack [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.335106] env[65385]: WARNING openstack [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.458769] env[65385]: DEBUG nova.network.neutron [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updated VIF entry in instance network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 789.459471] env[65385]: DEBUG nova.network.neutron [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 789.462119] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.368s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.465691] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.087s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.466124] env[65385]: DEBUG nova.objects.instance [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lazy-loading 'resources' on Instance uuid c798006c-3e74-461f-8397-3d8ca0be0a95 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 789.491279] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453634, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.496904] env[65385]: INFO nova.scheduler.client.report [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Deleted allocations for instance fae095a8-a768-4263-b456-ed0745398755 [ 789.515628] env[65385]: DEBUG nova.network.neutron [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Updated VIF entry in instance network info cache for port 6f259efd-d33d-4c33-8f8d-a98afc1ec5b8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 789.515812] env[65385]: DEBUG nova.network.neutron [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Updating instance_info_cache with network_info: [{"id": "7217d12a-b81f-4236-971b-a59c2709cf6f", "address": "fa:16:3e:c4:68:5c", "network": {"id": "4ff82148-f4e8-42a8-853a-d8d163ae26bd", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1544056117", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7217d12a-b8", "ovs_interfaceid": "7217d12a-b81f-4236-971b-a59c2709cf6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "address": "fa:16:3e:a6:d3:58", "network": {"id": "9159387d-e138-4204-8ed9-ac0e0da90550", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-474388752", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "21132a4099d84aaa89698915f7e91c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dabbac20-1723-40ad-9da0-e53b28073651", "external-id": "nsx-vlan-transportzone-790", "segmentation_id": 790, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f259efd-d3", "ovs_interfaceid": "6f259efd-d33d-4c33-8f8d-a98afc1ec5b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 789.558394] env[65385]: DEBUG nova.network.neutron [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Updating instance_info_cache with network_info: [{"id": "3ed3260d-d4bf-49a0-baa3-bb749653df5e", "address": "fa:16:3e:7b:61:84", "network": {"id": "fe370b84-e4c1-4975-a546-f721ab3fe25e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-632584528-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d75a2e244de46358945f6f2a57f1efc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ed3260d-d4", "ovs_interfaceid": "3ed3260d-d4bf-49a0-baa3-bb749653df5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 789.575873] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453635, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186356} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.576487] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.577681] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597ea2a5-2b17-4d36-bbf8-0045cd5762ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.587292] env[65385]: DEBUG oslo_vmware.api [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453636, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.614464] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] e9fd0c9f-5795-44df-9dcb-982b7722ae69/e9fd0c9f-5795-44df-9dcb-982b7722ae69.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.618687] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a49e2a06-2a7d-45a1-90cc-a5ee4a57e817 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.642638] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 789.642638] env[65385]: value = "task-4453638" [ 789.642638] env[65385]: _type = "Task" [ 789.642638] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.653725] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453638, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.675561] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453633, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.683305] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453637, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.739174] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba2a69d0-790e-4edf-9118-56b3d5ae364f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "c5d00bb5-d869-466f-826e-61622e57959f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.323s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.743991] env[65385]: WARNING openstack [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.747762] env[65385]: WARNING openstack [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.907134] env[65385]: WARNING neutronclient.v2_0.client [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.907829] env[65385]: WARNING openstack [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.908192] env[65385]: WARNING openstack [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.966514] env[65385]: DEBUG oslo_concurrency.lockutils [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] Releasing lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.967576] env[65385]: DEBUG nova.compute.manager [req-335c138e-3e11-42ab-b8f9-b1bae1d02e99 req-3cadd50b-6a74-4068-b047-fef9e9c5f0e8 service nova] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Received event network-vif-deleted-f18f6649-347b-43e6-9888-1dbb8af1227f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 789.967576] env[65385]: DEBUG oslo_concurrency.lockutils [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Acquired lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.967576] env[65385]: DEBUG nova.network.neutron [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Refreshing network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 789.996773] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453634, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.010213] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af17ded4-f3ce-4f71-96d4-c61b067bda48 tempest-ServerExternalEventsTest-902198954 tempest-ServerExternalEventsTest-902198954-project-member] Lock "fae095a8-a768-4263-b456-ed0745398755" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.249s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 790.022454] env[65385]: DEBUG oslo_concurrency.lockutils [req-e3d515bb-fca6-48d0-b1ce-e6daa2b79a14 req-9b1d62d0-fe31-4193-afc8-49716363b273 service nova] Releasing lock "refresh_cache-e9fd0c9f-5795-44df-9dcb-982b7722ae69" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.055045] env[65385]: DEBUG nova.network.neutron [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.064580] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Releasing lock "refresh_cache-ab8df1d8-615d-4083-bdf4-25f679a64982" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.065094] env[65385]: DEBUG nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Instance network_info: |[{"id": "3ed3260d-d4bf-49a0-baa3-bb749653df5e", "address": "fa:16:3e:7b:61:84", "network": {"id": "fe370b84-e4c1-4975-a546-f721ab3fe25e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-632584528-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d75a2e244de46358945f6f2a57f1efc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ed3260d-d4", "ovs_interfaceid": "3ed3260d-d4bf-49a0-baa3-bb749653df5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 790.065727] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:61:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aaf1b231-3660-4453-b4f3-44d825b9a5dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ed3260d-d4bf-49a0-baa3-bb749653df5e', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.075215] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Creating folder: Project (6d75a2e244de46358945f6f2a57f1efc). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.079794] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04b905ee-2b3b-4608-85b0-3662a11060a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.096850] env[65385]: DEBUG oslo_vmware.api [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453636, 'name': PowerOffVM_Task, 'duration_secs': 0.542427} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.100632] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 790.101148] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 790.102966] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5cba963-a79d-4aa5-a1a9-a2a23d030429 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.105041] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Created folder: Project (6d75a2e244de46358945f6f2a57f1efc) in parent group-v870881. [ 790.105234] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Creating folder: Instances. Parent ref: group-v871034. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.105889] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-281b8a9a-6f05-46d2-a66a-8f10ec81d945 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.119618] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Created folder: Instances in parent group-v871034. [ 790.119789] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 790.123683] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 790.126343] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7f381a2-4816-4063-abc0-0964143a0a37 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.157325] env[65385]: DEBUG nova.compute.manager [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Received event network-vif-plugged-3ed3260d-d4bf-49a0-baa3-bb749653df5e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 790.157325] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Acquiring lock "ab8df1d8-615d-4083-bdf4-25f679a64982-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 790.157325] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 790.157466] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 790.157791] env[65385]: DEBUG nova.compute.manager [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] No waiting events found dispatching network-vif-plugged-3ed3260d-d4bf-49a0-baa3-bb749653df5e {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 790.158021] env[65385]: WARNING nova.compute.manager [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Received unexpected event network-vif-plugged-3ed3260d-d4bf-49a0-baa3-bb749653df5e for instance with vm_state building and task_state spawning. [ 790.158236] env[65385]: DEBUG nova.compute.manager [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Received event network-changed-3ed3260d-d4bf-49a0-baa3-bb749653df5e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 790.158994] env[65385]: DEBUG nova.compute.manager [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Refreshing instance network info cache due to event network-changed-3ed3260d-d4bf-49a0-baa3-bb749653df5e. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 790.158994] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Acquiring lock "refresh_cache-ab8df1d8-615d-4083-bdf4-25f679a64982" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.158994] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Acquired lock "refresh_cache-ab8df1d8-615d-4083-bdf4-25f679a64982" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.159262] env[65385]: DEBUG nova.network.neutron [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Refreshing network info cache for port 3ed3260d-d4bf-49a0-baa3-bb749653df5e {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 790.163011] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.163011] env[65385]: value = "task-4453642" [ 790.163011] env[65385]: _type = "Task" [ 790.163011] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.179396] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453638, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.192958] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453633, 'name': CloneVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.201194] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453637, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.201465] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453642, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.239360] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 790.239597] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 790.239769] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Deleting the datastore file [datastore1] 2f66a03b-874b-4bee-9694-49813ecb0c8a {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.240201] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd72a347-738b-4fe6-817f-a1bbaca7955d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.250524] env[65385]: DEBUG oslo_vmware.api [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for the task: (returnval){ [ 790.250524] env[65385]: value = "task-4453643" [ 790.250524] env[65385]: _type = "Task" [ 790.250524] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.268309] env[65385]: DEBUG oslo_vmware.api [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453643, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.472653] env[65385]: WARNING neutronclient.v2_0.client [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.474631] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.474631] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.495599] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453634, 'name': CloneVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.560149] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.660926] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453638, 'name': ReconfigVM_Task, 'duration_secs': 0.53169} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.661259] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Reconfigured VM instance instance-00000034 to attach disk [datastore2] e9fd0c9f-5795-44df-9dcb-982b7722ae69/e9fd0c9f-5795-44df-9dcb-982b7722ae69.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.662090] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bbf6ea5f-4d14-4058-ae70-203d292b4d27 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.667216] env[65385]: WARNING neutronclient.v2_0.client [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.667881] env[65385]: WARNING openstack [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.668319] env[65385]: WARNING openstack [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.677039] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.677039] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.698732] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 790.698732] env[65385]: value = "task-4453644" [ 790.698732] env[65385]: _type = "Task" [ 790.698732] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.717023] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453633, 'name': CloneVM_Task, 'duration_secs': 1.614896} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.717023] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453642, 'name': CreateVM_Task, 'duration_secs': 0.513677} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.719144] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Created linked-clone VM from snapshot [ 790.719449] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 790.719755] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453637, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.720553] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3870fae3-4747-4c42-835c-907f6a5be360 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.724016] env[65385]: WARNING neutronclient.v2_0.client [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.724525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.724729] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.725147] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 790.737670] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0aa1c9a4-c119-46e2-9146-92f5d6946e9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.740038] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453644, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.747318] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 790.747318] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52537d70-9dc2-5774-1941-3246f0fee0fe" [ 790.747318] env[65385]: _type = "Task" [ 790.747318] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.752374] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Uploading image 30df6715-f055-4ee2-91ac-d0d74ee7d63d {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 790.763642] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91657f46-ea02-4145-abc6-b2feb3e4ae77 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.772436] env[65385]: DEBUG oslo_vmware.api [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Task: {'id': task-4453643, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268996} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.778543] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.778888] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 790.779111] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 790.779460] env[65385]: INFO nova.compute.manager [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Took 1.75 seconds to destroy the instance on the hypervisor. [ 790.779622] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 790.779885] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52537d70-9dc2-5774-1941-3246f0fee0fe, 'name': SearchDatastore_Task, 'duration_secs': 0.018381} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.781162] env[65385]: WARNING neutronclient.v2_0.client [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.781831] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.782725] env[65385]: WARNING openstack [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.792410] env[65385]: DEBUG nova.compute.manager [-] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 790.792553] env[65385]: DEBUG nova.network.neutron [-] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 790.792892] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.793433] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.793707] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.801857] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b4ee5c-2477-4b6c-a382-52388a5fa038 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.805594] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.805704] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.805885] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.806042] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.806277] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.807086] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f24901f-aaa0-4813-82e6-c75a841b2c18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.815517] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 790.815517] env[65385]: value = "vm-871032" [ 790.815517] env[65385]: _type = "VirtualMachine" [ 790.815517] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 790.815838] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-abbf2019-a92f-48fa-977b-22d60dafef63 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.852602] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.852838] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.855080] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aae24891-288a-4bdf-9d85-6192b55543b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.858614] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91796b2f-679d-46c2-a0e2-6d9518150959 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.867837] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lease: (returnval){ [ 790.867837] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b02f48-acc3-cd34-4858-e06534c76ed3" [ 790.867837] env[65385]: _type = "HttpNfcLease" [ 790.867837] env[65385]: } obtained for exporting VM: (result){ [ 790.867837] env[65385]: value = "vm-871032" [ 790.867837] env[65385]: _type = "VirtualMachine" [ 790.867837] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 790.868242] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the lease: (returnval){ [ 790.868242] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b02f48-acc3-cd34-4858-e06534c76ed3" [ 790.868242] env[65385]: _type = "HttpNfcLease" [ 790.868242] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 790.882353] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 790.882353] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fc1da-d4fa-4e42-e3d8-f65c0a864920" [ 790.882353] env[65385]: _type = "Task" [ 790.882353] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.885073] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.889094] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39deafe-3abe-4f87-baf0-733da46b31c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.895408] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 790.895408] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b02f48-acc3-cd34-4858-e06534c76ed3" [ 790.895408] env[65385]: _type = "HttpNfcLease" [ 790.895408] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 790.897964] env[65385]: WARNING openstack [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.899018] env[65385]: WARNING openstack [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.929185] env[65385]: DEBUG nova.compute.provider_tree [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.933107] env[65385]: DEBUG nova.network.neutron [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updated VIF entry in instance network info cache for port 6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 790.933461] env[65385]: DEBUG nova.network.neutron [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [{"id": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "address": "fa:16:3e:a0:70:59", "network": {"id": "9d7e0983-9bf3-40fa-bf56-e71eaea6ddbd", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-791168018-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e3431ecacb45288ccf0aba89cf5327", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c72dd42-3f", "ovs_interfaceid": "6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.938657] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523fc1da-d4fa-4e42-e3d8-f65c0a864920, 'name': SearchDatastore_Task, 'duration_secs': 0.04055} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.939513] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-487bdbe2-7ba7-4c88-aa62-52ce2b0fdde4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.948804] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 790.948804] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bcc614-95b4-0654-e5fa-764478a56bf8" [ 790.948804] env[65385]: _type = "Task" [ 790.948804] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.960690] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bcc614-95b4-0654-e5fa-764478a56bf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.993612] env[65385]: WARNING neutronclient.v2_0.client [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.994809] env[65385]: WARNING openstack [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.994809] env[65385]: WARNING openstack [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.008399] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453634, 'name': CloneVM_Task, 'duration_secs': 1.564727} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.010742] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Created linked-clone VM from snapshot [ 791.011930] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723ed192-21c5-4875-b29a-72e37e12c5b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.022963] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Uploading image 99c3648a-46c6-49a7-a427-b2729944e0d5 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 791.055233] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 791.055233] env[65385]: value = "vm-871033" [ 791.055233] env[65385]: _type = "VirtualMachine" [ 791.055233] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 791.055460] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-000d9c4f-26fb-4931-bd0c-6a3ebc9e73f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.065545] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease: (returnval){ [ 791.065545] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c84fff-fa27-f862-ed8d-2be1ad5ba53e" [ 791.065545] env[65385]: _type = "HttpNfcLease" [ 791.065545] env[65385]: } obtained for exporting VM: (result){ [ 791.065545] env[65385]: value = "vm-871033" [ 791.065545] env[65385]: _type = "VirtualMachine" [ 791.065545] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 791.065794] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the lease: (returnval){ [ 791.065794] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c84fff-fa27-f862-ed8d-2be1ad5ba53e" [ 791.065794] env[65385]: _type = "HttpNfcLease" [ 791.065794] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 791.066135] env[65385]: DEBUG nova.compute.manager [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 791.067032] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35849b73-b370-4273-b922-d669a348b7c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.082229] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 791.082229] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c84fff-fa27-f862-ed8d-2be1ad5ba53e" [ 791.082229] env[65385]: _type = "HttpNfcLease" [ 791.082229] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 791.175391] env[65385]: DEBUG nova.network.neutron [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Updated VIF entry in instance network info cache for port 3ed3260d-d4bf-49a0-baa3-bb749653df5e. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 791.175391] env[65385]: DEBUG nova.network.neutron [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Updating instance_info_cache with network_info: [{"id": "3ed3260d-d4bf-49a0-baa3-bb749653df5e", "address": "fa:16:3e:7b:61:84", "network": {"id": "fe370b84-e4c1-4975-a546-f721ab3fe25e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-632584528-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d75a2e244de46358945f6f2a57f1efc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aaf1b231-3660-4453-b4f3-44d825b9a5dd", "external-id": "nsx-vlan-transportzone-6", "segmentation_id": 6, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ed3260d-d4", "ovs_interfaceid": "3ed3260d-d4bf-49a0-baa3-bb749653df5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 791.204352] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453637, 'name': ReconfigVM_Task, 'duration_secs': 1.569097} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.210370] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.210715] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance '38e0ebf5-1d92-4906-b0bd-b0ebf8651f22' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 791.227887] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453644, 'name': Rename_Task, 'duration_secs': 0.188747} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.229518] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.229518] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ddd9362f-9c04-42ca-995d-5f605ab63887 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.239891] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 791.239891] env[65385]: value = "task-4453647" [ 791.239891] env[65385]: _type = "Task" [ 791.239891] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.252736] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.385089] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 791.385089] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b02f48-acc3-cd34-4858-e06534c76ed3" [ 791.385089] env[65385]: _type = "HttpNfcLease" [ 791.385089] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 791.386174] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 791.386174] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b02f48-acc3-cd34-4858-e06534c76ed3" [ 791.386174] env[65385]: _type = "HttpNfcLease" [ 791.386174] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 791.390029] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef38b0e7-aa37-4ebd-998c-fd4e7ba90a59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.404020] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e2420a-4793-f63e-f2dc-892ea4528440/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 791.404020] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e2420a-4793-f63e-f2dc-892ea4528440/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 791.470732] env[65385]: DEBUG nova.scheduler.client.report [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 791.474771] env[65385]: DEBUG oslo_concurrency.lockutils [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] Releasing lock "refresh_cache-2f66a03b-874b-4bee-9694-49813ecb0c8a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.475178] env[65385]: DEBUG nova.compute.manager [req-eeb03e95-7cde-4449-a3cd-32606a3d16dd req-48bcaea4-edb2-4fee-bc45-2b2d930485a7 service nova] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Received event network-vif-deleted-7a4f4b0b-f75e-4ffe-bb2b-a52fc81af719 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 791.492555] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bcc614-95b4-0654-e5fa-764478a56bf8, 'name': SearchDatastore_Task, 'duration_secs': 0.023131} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.493570] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.493962] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] ab8df1d8-615d-4083-bdf4-25f679a64982/ab8df1d8-615d-4083-bdf4-25f679a64982.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.494293] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-526337e1-817e-4500-a39f-22ccca43db19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.504050] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 791.504050] env[65385]: value = "task-4453648" [ 791.504050] env[65385]: _type = "Task" [ 791.504050] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.515575] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453648, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.520542] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c28cb2-db14-49da-b488-2b93fb72d0a1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.529881] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55887644-8e7a-463a-83b0-7731a9a67874 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Suspending the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 791.530319] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-554649ef-cf15-429c-9912-5198e73c7952 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.536600] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d2c66f25-04b5-4c96-9caf-1b96df4b52c5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.540720] env[65385]: DEBUG oslo_vmware.api [None req-55887644-8e7a-463a-83b0-7731a9a67874 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 791.540720] env[65385]: value = "task-4453649" [ 791.540720] env[65385]: _type = "Task" [ 791.540720] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.553026] env[65385]: DEBUG oslo_vmware.api [None req-55887644-8e7a-463a-83b0-7731a9a67874 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453649, 'name': SuspendVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.581245] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 791.581245] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c84fff-fa27-f862-ed8d-2be1ad5ba53e" [ 791.581245] env[65385]: _type = "HttpNfcLease" [ 791.581245] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 791.581921] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 791.581921] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c84fff-fa27-f862-ed8d-2be1ad5ba53e" [ 791.581921] env[65385]: _type = "HttpNfcLease" [ 791.581921] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 791.582847] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f8546c-05ba-4d0c-9b59-4c5fda3fd124 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.595996] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52430440-fc66-b7c2-8f09-27d29d76d815/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 791.595996] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52430440-fc66-b7c2-8f09-27d29d76d815/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 791.679719] env[65385]: DEBUG oslo_concurrency.lockutils [req-d6f316ea-6fff-4430-bc92-2b98de231cfc req-c93b6fc0-9b73-4994-8e27-c3f60bbf5f96 service nova] Releasing lock "refresh_cache-ab8df1d8-615d-4083-bdf4-25f679a64982" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.722312] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc46d52-a987-41ab-a0eb-01da89332052 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.725928] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6527f5ed-65e6-4431-aa74-c43b4da565e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.757593] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c827e8-4ce1-4fde-9a4f-4dd588dfe41c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.772929] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453647, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.797890] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance '38e0ebf5-1d92-4906-b0bd-b0ebf8651f22' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 791.980029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.514s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.983942] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.815s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.985931] env[65385]: INFO nova.compute.claims [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.024280] env[65385]: DEBUG nova.network.neutron [-] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 792.036520] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453648, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.067398] env[65385]: DEBUG oslo_vmware.api [None req-55887644-8e7a-463a-83b0-7731a9a67874 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453649, 'name': SuspendVM_Task} progress is 58%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.071035] env[65385]: INFO nova.scheduler.client.report [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Deleted allocations for instance c798006c-3e74-461f-8397-3d8ca0be0a95 [ 792.092199] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67174106-30ca-469f-8c37-7ca6b1d7d183 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.105207] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Doing hard reboot of VM {{(pid=65385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 792.105369] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-98d56aac-6835-4561-b297-5e079e44c8c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.118018] env[65385]: DEBUG oslo_vmware.api [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 792.118018] env[65385]: value = "task-4453650" [ 792.118018] env[65385]: _type = "Task" [ 792.118018] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.135632] env[65385]: DEBUG oslo_vmware.api [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453650, 'name': ResetVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.204649] env[65385]: DEBUG nova.compute.manager [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 792.206290] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466f9327-4801-40c9-a146-024d35a6da14 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.265435] env[65385]: DEBUG oslo_vmware.api [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453647, 'name': PowerOnVM_Task, 'duration_secs': 0.660412} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.265620] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.266035] env[65385]: INFO nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Took 11.32 seconds to spawn the instance on the hypervisor. [ 792.266364] env[65385]: DEBUG nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 792.267542] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485f1363-5d15-42ab-8b97-10495ae76a3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.310026] env[65385]: WARNING neutronclient.v2_0.client [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.311817] env[65385]: WARNING neutronclient.v2_0.client [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.442948] env[65385]: DEBUG nova.network.neutron [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Port ad95d76f-e023-47b2-a936-e7ed11779276 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 792.535493] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453648, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654859} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.536820] env[65385]: INFO nova.compute.manager [-] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Took 1.74 seconds to deallocate network for instance. [ 792.537928] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] ab8df1d8-615d-4083-bdf4-25f679a64982/ab8df1d8-615d-4083-bdf4-25f679a64982.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.541809] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.545433] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f43f8559-1640-42b8-91a7-4b21d4bf2115 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.573871] env[65385]: DEBUG oslo_vmware.api [None req-55887644-8e7a-463a-83b0-7731a9a67874 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453649, 'name': SuspendVM_Task, 'duration_secs': 0.818638} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.578468] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55887644-8e7a-463a-83b0-7731a9a67874 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Suspended the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 792.578877] env[65385]: DEBUG nova.compute.manager [None req-55887644-8e7a-463a-83b0-7731a9a67874 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 792.579219] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 792.579219] env[65385]: value = "task-4453651" [ 792.579219] env[65385]: _type = "Task" [ 792.579219] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.590191] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40ddb4f-bfdd-4cea-b80b-baf58c2c3607 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.597905] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8ab731c5-7a4a-44c3-ba67-ba8a5eaa16da tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "c798006c-3e74-461f-8397-3d8ca0be0a95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.866s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.632854] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453651, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.645202] env[65385]: DEBUG oslo_vmware.api [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453650, 'name': ResetVM_Task, 'duration_secs': 0.13695} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.645644] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Did hard reboot of VM {{(pid=65385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 792.645860] env[65385]: DEBUG nova.compute.manager [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 792.647190] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa80355-8701-4088-894f-10aa738a5330 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.705883] env[65385]: DEBUG nova.compute.manager [req-9a35ac00-34be-4e2d-bdaf-fd9a7b617910 req-c3bd2852-c9d1-419e-98d4-3b6f0641d79d service nova] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Received event network-vif-deleted-6c72dd42-3f1f-4b31-a0b1-8c6a7b3e0274 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 792.711761] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "34730159-3bec-4ad5-b85e-0f67998d6001" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.713269] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "34730159-3bec-4ad5-b85e-0f67998d6001" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.713723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "34730159-3bec-4ad5-b85e-0f67998d6001-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.714010] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "34730159-3bec-4ad5-b85e-0f67998d6001-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.714162] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "34730159-3bec-4ad5-b85e-0f67998d6001-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.719937] env[65385]: INFO nova.compute.manager [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Terminating instance [ 792.730647] env[65385]: INFO nova.compute.manager [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] instance snapshotting [ 792.744030] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca1dfdd-76de-4d8f-b251-5e7886f46db6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.776030] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388f5bdb-07b7-4722-bca3-88100ab33a87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.804664] env[65385]: INFO nova.compute.manager [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Took 46.15 seconds to build instance. [ 793.055598] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.107077] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453651, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123089} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.107418] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.108290] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb231de-f899-4ca4-bcf0-6462def3186d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.147798] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] ab8df1d8-615d-4083-bdf4-25f679a64982/ab8df1d8-615d-4083-bdf4-25f679a64982.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.153221] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8884d1e6-1c3e-41d9-8cdc-e650891cf323 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.179154] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7c1c6784-3f47-43c5-a61e-da1f6c1097f3 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.884s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.184579] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 793.184579] env[65385]: value = "task-4453652" [ 793.184579] env[65385]: _type = "Task" [ 793.184579] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.199580] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453652, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.242037] env[65385]: DEBUG nova.compute.manager [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 793.242156] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.244623] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11cf32b-27aa-43f7-8e57-97a45bb9ccda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.260356] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 793.260735] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57a8fa86-b885-4965-be34-aaf66542ab0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.269644] env[65385]: DEBUG oslo_vmware.api [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 793.269644] env[65385]: value = "task-4453653" [ 793.269644] env[65385]: _type = "Task" [ 793.269644] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.286647] env[65385]: DEBUG oslo_vmware.api [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.307306] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 793.311081] env[65385]: DEBUG oslo_concurrency.lockutils [None req-54447323-3b54-4720-8d49-f98a7c2bbb7a tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.613s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.311518] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-346d6d4d-5bd9-4d95-9803-8db46143f00b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.322698] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 793.322698] env[65385]: value = "task-4453654" [ 793.322698] env[65385]: _type = "Task" [ 793.322698] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.342771] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453654, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.493340] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.494051] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.004s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.494843] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.704934] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.762186] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebbf3d5-b378-4080-8eed-a9454b0acb9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.779401] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b59b1c-70b9-44be-94e6-047566e9f5b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.793392] env[65385]: DEBUG oslo_vmware.api [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.829966] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25aae183-499d-4d1f-90ac-f7ab4197b60b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.840052] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453654, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.844138] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afcda71-61be-47c1-ad02-4fa0a87dffcf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.866355] env[65385]: DEBUG nova.compute.provider_tree [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.197101] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453652, 'name': ReconfigVM_Task, 'duration_secs': 0.974263} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.197418] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Reconfigured VM instance instance-00000035 to attach disk [datastore2] ab8df1d8-615d-4083-bdf4-25f679a64982/ab8df1d8-615d-4083-bdf4-25f679a64982.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.198145] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9201e261-7fb4-44e4-a192-067a4a13dd5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.207205] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 794.207205] env[65385]: value = "task-4453655" [ 794.207205] env[65385]: _type = "Task" [ 794.207205] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.221571] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453655, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.283731] env[65385]: DEBUG oslo_vmware.api [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453653, 'name': PowerOffVM_Task, 'duration_secs': 0.784918} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.284057] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 794.284233] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 794.284541] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ef1481e-f574-4dd3-b5a4-db6336a32d7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.339236] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453654, 'name': CreateSnapshot_Task, 'duration_secs': 0.839438} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.339397] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 794.340186] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b667c7-61f2-40db-8d94-204701d574a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.372481] env[65385]: DEBUG nova.scheduler.client.report [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 794.378411] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 794.379336] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 794.379336] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleting the datastore file [datastore2] 34730159-3bec-4ad5-b85e-0f67998d6001 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.379940] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96d8edf4-93a9-4e97-a168-999b49c2b020 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.390428] env[65385]: DEBUG oslo_vmware.api [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 794.390428] env[65385]: value = "task-4453657" [ 794.390428] env[65385]: _type = "Task" [ 794.390428] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.403404] env[65385]: DEBUG oslo_vmware.api [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453657, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.510467] env[65385]: WARNING neutronclient.v2_0.client [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.559062] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.559251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.559434] env[65385]: DEBUG nova.network.neutron [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 794.719899] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453655, 'name': Rename_Task, 'duration_secs': 0.207119} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.720306] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.720381] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed80b2f8-e180-4519-a403-3677efc28280 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.729387] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 794.729387] env[65385]: value = "task-4453658" [ 794.729387] env[65385]: _type = "Task" [ 794.729387] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.740371] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453658, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.862717] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 794.863077] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a5627228-dd0f-4ed9-b843-e5d11c66b3d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.872996] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 794.872996] env[65385]: value = "task-4453659" [ 794.872996] env[65385]: _type = "Task" [ 794.872996] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.883871] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.900s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.884631] env[65385]: DEBUG nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 794.888953] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453659, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.889406] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.283s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 794.889662] env[65385]: DEBUG nova.objects.instance [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 794.903574] env[65385]: DEBUG oslo_vmware.api [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453657, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.312885} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.904018] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.904873] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.904873] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.904873] env[65385]: INFO nova.compute.manager [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Took 1.66 seconds to destroy the instance on the hypervisor. [ 794.904873] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 794.905173] env[65385]: DEBUG nova.compute.manager [-] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 794.905173] env[65385]: DEBUG nova.network.neutron [-] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 794.905423] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.906087] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.906361] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.991088] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.062393] env[65385]: WARNING neutronclient.v2_0.client [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.063366] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.063765] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.246201] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453658, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.318299] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.318299] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.389459] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453659, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.395513] env[65385]: DEBUG nova.compute.utils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 795.401934] env[65385]: WARNING neutronclient.v2_0.client [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.402682] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.403128] env[65385]: WARNING openstack [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.412204] env[65385]: DEBUG nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 795.510333] env[65385]: DEBUG nova.network.neutron [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance_info_cache with network_info: [{"id": "ad95d76f-e023-47b2-a936-e7ed11779276", "address": "fa:16:3e:b0:c7:d2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95d76f-e0", "ovs_interfaceid": "ad95d76f-e023-47b2-a936-e7ed11779276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.747032] env[65385]: DEBUG nova.network.neutron [-] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.750261] env[65385]: DEBUG oslo_vmware.api [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453658, 'name': PowerOnVM_Task, 'duration_secs': 0.968188} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.752954] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.752954] env[65385]: INFO nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Took 8.97 seconds to spawn the instance on the hypervisor. [ 795.752954] env[65385]: DEBUG nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 795.753928] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c922889-3944-4584-9fe8-8311809adba1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.887096] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453659, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.915033] env[65385]: DEBUG nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 795.919092] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdc820cb-c910-4dee-b485-a650a8caec5d tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.028s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.919092] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.783s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.920724] env[65385]: INFO nova.compute.claims [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.013016] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 796.252579] env[65385]: INFO nova.compute.manager [-] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Took 1.35 seconds to deallocate network for instance. [ 796.276023] env[65385]: INFO nova.compute.manager [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Took 39.43 seconds to build instance. [ 796.387501] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453659, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.552799] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f5d07e-7df6-4a29-81bd-7de051244278 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.573869] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b4ae22-59c4-4252-b263-79fe324efb9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.582724] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance '38e0ebf5-1d92-4906-b0bd-b0ebf8651f22' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 796.759776] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.779009] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a523e6e9-0505-4ab8-bcb5-0484d025f493 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.772s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.890825] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453659, 'name': CloneVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.924891] env[65385]: DEBUG nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 797.090630] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 797.090959] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f2533f9-196f-44e5-9760-a855abb09b3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.099990] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 797.099990] env[65385]: value = "task-4453660" [ 797.099990] env[65385]: _type = "Task" [ 797.099990] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.109456] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.389339] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453659, 'name': CloneVM_Task, 'duration_secs': 2.127848} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.389682] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Created linked-clone VM from snapshot [ 797.390481] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c2a5a2-da33-456c-83ed-0c8b86e70050 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.400513] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Uploading image 71310ad2-3ac2-4dcf-9d77-a75979646daf {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 797.467673] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1663cdca-c916-4e56-b900-e3f4a8ffd301 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.478485] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e678785e-1441-453d-b56f-faaa670c2853 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.513046] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6503e95-3255-4c93-8b67-3e682e7a3e31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.522646] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468d97f7-6340-40ec-9b04-8e38e8a87167 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.539910] env[65385]: DEBUG nova.compute.provider_tree [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.610595] env[65385]: DEBUG oslo_vmware.api [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453660, 'name': PowerOnVM_Task, 'duration_secs': 0.48674} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.610869] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 797.611419] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff43b60-58d5-491c-861a-06ddf1978e9d tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance '38e0ebf5-1d92-4906-b0bd-b0ebf8651f22' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 798.044169] env[65385]: DEBUG nova.scheduler.client.report [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 798.552908] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.552908] env[65385]: DEBUG nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 798.555097] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 30.184s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.555585] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.555913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 798.556326] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.952s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.556702] env[65385]: DEBUG nova.objects.instance [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 798.561249] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7d22da-4e48-4002-9298-788162dfbe79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.571922] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2880f27-4cad-406f-aecf-da8ee1499046 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.590866] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b41c009-cd67-4892-a697-03f42c809657 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.601906] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfc9389-b282-4a80-895c-b3e0d707d135 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.638848] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177813MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 798.639074] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.062329] env[65385]: DEBUG nova.compute.utils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 799.066796] env[65385]: DEBUG nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 799.067048] env[65385]: DEBUG nova.network.neutron [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 799.067685] env[65385]: WARNING neutronclient.v2_0.client [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.067790] env[65385]: WARNING neutronclient.v2_0.client [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.068537] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.069071] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.079102] env[65385]: DEBUG oslo_concurrency.lockutils [None req-626dc079-9342-4007-8f3e-6876704ff069 tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.523s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.080575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.865s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.080889] env[65385]: DEBUG nova.objects.instance [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lazy-loading 'resources' on Instance uuid 53c31e42-4db0-4428-8336-b3a82e3f0e8f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 799.133039] env[65385]: DEBUG nova.policy [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03f7c78b16e547c38f24ed2c8a329202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ba61d77cc2d4fe1ba3c03466d5985d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 799.490807] env[65385]: DEBUG nova.network.neutron [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Successfully created port: 622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 799.571558] env[65385]: DEBUG nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 800.120106] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd55524e-9f77-498c-992c-5dded7b87746 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.130435] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1cc48a-567e-483a-8faa-27629817f77d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.163664] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb9ef82-50a6-4e63-8942-8c340ce1c226 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.172363] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3654bd-e91f-497f-90b3-2f77c45bd2f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.188224] env[65385]: DEBUG nova.compute.provider_tree [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.580015] env[65385]: DEBUG nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 800.692080] env[65385]: DEBUG nova.scheduler.client.report [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 801.170533] env[65385]: DEBUG nova.network.neutron [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Successfully updated port: 622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 801.197820] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.117s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.200743] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.941s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.201908] env[65385]: INFO nova.compute.claims [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.223797] env[65385]: INFO nova.scheduler.client.report [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Deleted allocations for instance 53c31e42-4db0-4428-8336-b3a82e3f0e8f [ 801.675704] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.675947] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 801.676151] env[65385]: DEBUG nova.network.neutron [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 801.735309] env[65385]: DEBUG oslo_concurrency.lockutils [None req-07eb4f03-682c-42f1-a4b6-8cd04f9ab1dc tempest-ServersListShow2100Test-1819160414 tempest-ServersListShow2100Test-1819160414-project-member] Lock "53c31e42-4db0-4428-8336-b3a82e3f0e8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.381s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.869936] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 801.870366] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-23aa6a1b-c8ae-4fcd-8a8f-faeab5e96bf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.880244] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 801.880244] env[65385]: value = "task-4453661" [ 801.880244] env[65385]: _type = "Task" [ 801.880244] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.891365] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453661, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.907811] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 801.908497] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.908847] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 801.909359] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.909716] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 801.910295] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 801.910524] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.910888] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 801.911313] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 801.911618] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 801.911897] env[65385]: DEBUG nova.virt.hardware [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 801.915434] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cffeb4-05ee-4c79-9501-79e7391382d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.925801] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96341ad9-0eb5-41ee-95a1-1766eeb85089 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.934181] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 801.934181] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.934412] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 801.934412] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.934988] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 801.934988] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 801.934988] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.935296] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 801.935296] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 801.935436] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 801.935594] env[65385]: DEBUG nova.virt.hardware [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 801.937852] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cc5383-8a31-4819-bde4-9e3139f24574 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.954902] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.960457] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Creating folder: Project (04be7e71aa654488a30f2a7afccc9f31). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.963175] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52430440-fc66-b7c2-8f09-27d29d76d815/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 801.963575] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce645afd-8820-4f95-a550-8b7078749378 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.966486] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e04c6c-2370-4de6-a841-bd3826c9fe1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.971313] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d695a2e4-3e83-4a4e-95f0-8beceaad83b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.989176] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Created folder: Project (04be7e71aa654488a30f2a7afccc9f31) in parent group-v870881. [ 801.989410] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Creating folder: Instances. Parent ref: group-v871039. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.989708] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52430440-fc66-b7c2-8f09-27d29d76d815/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 801.989859] env[65385]: ERROR oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52430440-fc66-b7c2-8f09-27d29d76d815/disk-0.vmdk due to incomplete transfer. [ 801.990113] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c0f01a9d-60a4-42d7-b33f-3d9867a79ac3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.991885] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9e0a3b48-863f-4ebb-b57e-530bfcfd9138 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.001130] env[65385]: DEBUG oslo_vmware.rw_handles [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52430440-fc66-b7c2-8f09-27d29d76d815/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 802.001340] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Uploaded image 99c3648a-46c6-49a7-a427-b2729944e0d5 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 802.003071] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 802.004719] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e78c4d83-d1ed-4cb7-a6df-48792a0e31f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.006612] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Created folder: Instances in parent group-v871039. [ 802.006882] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 802.007120] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.007793] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0f22531-e986-4ac9-96d4-f084643bfd50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.022627] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 802.022627] env[65385]: value = "task-4453664" [ 802.022627] env[65385]: _type = "Task" [ 802.022627] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.028955] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.028955] env[65385]: value = "task-4453665" [ 802.028955] env[65385]: _type = "Task" [ 802.028955] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.038243] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453664, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.044770] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453665, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.182165] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.182749] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.255331] env[65385]: DEBUG nova.network.neutron [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 802.353234] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e2420a-4793-f63e-f2dc-892ea4528440/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 802.354720] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011359f9-9029-440e-b7a4-894f0898e1c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.367206] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e2420a-4793-f63e-f2dc-892ea4528440/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 802.367206] env[65385]: ERROR oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e2420a-4793-f63e-f2dc-892ea4528440/disk-0.vmdk due to incomplete transfer. [ 802.367206] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5bd8311e-26fd-4aa9-8599-8a6b484e58da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.370658] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.371120] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.394644] env[65385]: DEBUG oslo_vmware.rw_handles [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e2420a-4793-f63e-f2dc-892ea4528440/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 802.394923] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Uploaded image 30df6715-f055-4ee2-91ac-d0d74ee7d63d to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 802.400061] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 802.407696] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1ef7c918-a73f-4c18-b190-98b79f132b38 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.418324] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453661, 'name': Destroy_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.419965] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 802.419965] env[65385]: value = "task-4453666" [ 802.419965] env[65385]: _type = "Task" [ 802.419965] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.430617] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453666, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.496734] env[65385]: WARNING neutronclient.v2_0.client [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.497591] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.497987] env[65385]: WARNING openstack [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.539198] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453664, 'name': Destroy_Task, 'duration_secs': 0.451771} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.545082] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Destroyed the VM [ 802.545345] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 802.545582] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453665, 'name': CreateVM_Task, 'duration_secs': 0.493195} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.545999] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c67615e7-844e-4395-82c1-f04d80179809 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.547876] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.548315] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.548543] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.548847] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 802.549473] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6ff527d-57e4-45f4-942d-b26625d930dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.563079] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 802.563079] env[65385]: value = "task-4453667" [ 802.563079] env[65385]: _type = "Task" [ 802.563079] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.563381] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 802.563381] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f063dd-31a7-e2d9-4d90-0807b62678a9" [ 802.563381] env[65385]: _type = "Task" [ 802.563381] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.581089] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f063dd-31a7-e2d9-4d90-0807b62678a9, 'name': SearchDatastore_Task, 'duration_secs': 0.015584} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.581436] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453667, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.581746] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 802.581965] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.582215] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.582364] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.582608] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.582973] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bd46f49-04d3-4d36-8193-c148e8bb198f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.595980] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.596208] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.596999] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15c1a075-1f24-47f9-985f-f10a8c2fab87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.603564] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 802.603564] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521b283b-7729-76a6-0e67-2fbb5807601b" [ 802.603564] env[65385]: _type = "Task" [ 802.603564] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.613537] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521b283b-7729-76a6-0e67-2fbb5807601b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.627722] env[65385]: DEBUG nova.network.neutron [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Updating instance_info_cache with network_info: [{"id": "622e8c02-7c83-4b27-9884-f6c399712a84", "address": "fa:16:3e:a7:44:e2", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap622e8c02-7c", "ovs_interfaceid": "622e8c02-7c83-4b27-9884-f6c399712a84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.828191] env[65385]: DEBUG nova.compute.manager [req-31dc67f0-1358-447f-b9c7-766f00b23bec req-f967eaea-37ca-42ed-98ab-67e641733a66 service nova] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Received event network-vif-deleted-884a69db-a84a-46ae-9af6-bd5f21a386e7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 802.858882] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b87a500-f69b-4c83-86da-68a565aa3dd5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.867898] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e6f346-7552-4dcd-aeaa-44b01d4caee4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.875557] env[65385]: DEBUG nova.compute.manager [req-1c9dd840-966b-438d-9ec3-926128638cb7 req-0263ac10-9464-4eac-a774-520685c32b03 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Received event network-vif-plugged-622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 802.875557] env[65385]: DEBUG oslo_concurrency.lockutils [req-1c9dd840-966b-438d-9ec3-926128638cb7 req-0263ac10-9464-4eac-a774-520685c32b03 service nova] Acquiring lock "26a2d464-0aab-463f-8265-948a4f0ff188-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.875767] env[65385]: DEBUG oslo_concurrency.lockutils [req-1c9dd840-966b-438d-9ec3-926128638cb7 req-0263ac10-9464-4eac-a774-520685c32b03 service nova] Lock "26a2d464-0aab-463f-8265-948a4f0ff188-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.876084] env[65385]: DEBUG oslo_concurrency.lockutils [req-1c9dd840-966b-438d-9ec3-926128638cb7 req-0263ac10-9464-4eac-a774-520685c32b03 service nova] Lock "26a2d464-0aab-463f-8265-948a4f0ff188-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.876302] env[65385]: DEBUG nova.compute.manager [req-1c9dd840-966b-438d-9ec3-926128638cb7 req-0263ac10-9464-4eac-a774-520685c32b03 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] No waiting events found dispatching network-vif-plugged-622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 802.876975] env[65385]: WARNING nova.compute.manager [req-1c9dd840-966b-438d-9ec3-926128638cb7 req-0263ac10-9464-4eac-a774-520685c32b03 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Received unexpected event network-vif-plugged-622e8c02-7c83-4b27-9884-f6c399712a84 for instance with vm_state building and task_state spawning. [ 802.910323] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316c43e7-816f-478c-b458-4d41851db3a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.920649] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453661, 'name': Destroy_Task, 'duration_secs': 0.539033} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.922912] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Destroyed the VM [ 802.923268] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 802.926720] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6849643e-0f32-40e2-ab80-88f2056b9835 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.930526] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e68feb-756b-4867-958a-e80a28a86c02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.951666] env[65385]: DEBUG nova.compute.provider_tree [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.954399] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453666, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.956571] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 802.956571] env[65385]: value = "task-4453668" [ 802.956571] env[65385]: _type = "Task" [ 802.956571] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.966012] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453668, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.076206] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453667, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.116546] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521b283b-7729-76a6-0e67-2fbb5807601b, 'name': SearchDatastore_Task, 'duration_secs': 0.011124} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.117467] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a32ca30a-712b-4763-9dfd-ec95e5ecb9ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.125776] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 803.125776] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520ad049-82d1-3811-9229-ba4b5d15d80d" [ 803.125776] env[65385]: _type = "Task" [ 803.125776] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.132404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.133064] env[65385]: DEBUG nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Instance network_info: |[{"id": "622e8c02-7c83-4b27-9884-f6c399712a84", "address": "fa:16:3e:a7:44:e2", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap622e8c02-7c", "ovs_interfaceid": "622e8c02-7c83-4b27-9884-f6c399712a84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 803.133339] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:44:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '622e8c02-7c83-4b27-9884-f6c399712a84', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.141601] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Creating folder: Project (2ba61d77cc2d4fe1ba3c03466d5985d9). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.142186] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07d9bb33-345e-4b98-acaf-5a79fe61e6b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.148625] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520ad049-82d1-3811-9229-ba4b5d15d80d, 'name': SearchDatastore_Task, 'duration_secs': 0.011807} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.148912] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.149201] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 65f9bc55-7bab-43b0-a974-eb5080389b7d/65f9bc55-7bab-43b0-a974-eb5080389b7d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.149513] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32f8347b-30c3-4ff8-86ba-484e6bd7a990 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.156158] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Created folder: Project (2ba61d77cc2d4fe1ba3c03466d5985d9) in parent group-v870881. [ 803.156453] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Creating folder: Instances. Parent ref: group-v871042. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.156839] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae020733-709c-47a6-87a8-c3346b8d3884 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.160950] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 803.160950] env[65385]: value = "task-4453670" [ 803.160950] env[65385]: _type = "Task" [ 803.160950] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.172114] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.173988] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Created folder: Instances in parent group-v871042. [ 803.174660] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 803.174660] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 803.174872] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db6e2f9c-4d44-4d05-a719-380bfcbee6d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.203040] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.203040] env[65385]: value = "task-4453672" [ 803.203040] env[65385]: _type = "Task" [ 803.203040] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.219502] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453672, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.435857] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453666, 'name': Destroy_Task, 'duration_secs': 0.809776} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.436180] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Destroyed the VM [ 803.436651] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 803.436934] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cc1d432a-7883-4db3-8a87-88179215ba33 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.445572] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 803.445572] env[65385]: value = "task-4453673" [ 803.445572] env[65385]: _type = "Task" [ 803.445572] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.457049] env[65385]: DEBUG nova.scheduler.client.report [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 803.461845] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453673, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.481926] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453668, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.576881] env[65385]: DEBUG oslo_vmware.api [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453667, 'name': RemoveSnapshot_Task, 'duration_secs': 0.67265} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.577208] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 803.577528] env[65385]: INFO nova.compute.manager [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Took 16.81 seconds to snapshot the instance on the hypervisor. [ 803.676287] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453670, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.717269] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453672, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.795363] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.796138] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.796138] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.796233] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.796471] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.799395] env[65385]: INFO nova.compute.manager [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Terminating instance [ 803.956386] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453673, 'name': RemoveSnapshot_Task} progress is 74%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.970468] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.768s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.971057] env[65385]: DEBUG nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 803.978783] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.939s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.980305] env[65385]: INFO nova.compute.claims [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.985473] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453668, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.991174] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.131302] env[65385]: DEBUG nova.compute.manager [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Found 3 images (rotation: 2) {{(pid=65385) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 804.131511] env[65385]: DEBUG nova.compute.manager [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Rotating out 1 backups {{(pid=65385) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5065}} [ 804.131681] env[65385]: DEBUG nova.compute.manager [None req-7707c8ed-fe6a-49df-87da-24bd5cf527ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleting image 7b43aa8d-6c77-4694-9308-fdf00d082211 {{(pid=65385) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5070}} [ 804.147733] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.148152] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.148414] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.172948] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453670, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644775} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.173210] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 65f9bc55-7bab-43b0-a974-eb5080389b7d/65f9bc55-7bab-43b0-a974-eb5080389b7d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.173505] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.173848] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2af91db-8d3c-4414-925a-8cfbde72fce9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.182319] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 804.182319] env[65385]: value = "task-4453674" [ 804.182319] env[65385]: _type = "Task" [ 804.182319] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.191645] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.202780] env[65385]: DEBUG nova.network.neutron [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Port ad95d76f-e023-47b2-a936-e7ed11779276 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 804.202780] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.202940] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.203050] env[65385]: DEBUG nova.network.neutron [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 804.214232] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453672, 'name': CreateVM_Task, 'duration_secs': 0.714281} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.215134] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.215585] env[65385]: WARNING neutronclient.v2_0.client [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.215949] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.216124] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.216468] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 804.216995] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0be0fcd-6907-45d8-b15d-d64a5815b133 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.223561] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 804.223561] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c478db-93f1-2fe0-5467-ee036a5fc5b0" [ 804.223561] env[65385]: _type = "Task" [ 804.223561] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.235231] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c478db-93f1-2fe0-5467-ee036a5fc5b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.306329] env[65385]: DEBUG nova.compute.manager [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 804.306800] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.308049] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc71b40-15c2-48ef-9234-f37d8306b78f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.318016] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.321020] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23c3ec8a-2391-4ec9-ae64-de78c0f639c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.326153] env[65385]: DEBUG oslo_vmware.api [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 804.326153] env[65385]: value = "task-4453675" [ 804.326153] env[65385]: _type = "Task" [ 804.326153] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.336238] env[65385]: DEBUG oslo_vmware.api [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453675, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.459177] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453673, 'name': RemoveSnapshot_Task} progress is 74%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.474189] env[65385]: DEBUG oslo_vmware.api [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453668, 'name': RemoveSnapshot_Task, 'duration_secs': 1.104486} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.474520] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 804.491277] env[65385]: DEBUG nova.compute.utils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 804.494350] env[65385]: DEBUG nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 804.494350] env[65385]: DEBUG nova.network.neutron [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 804.494350] env[65385]: WARNING neutronclient.v2_0.client [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.494350] env[65385]: WARNING neutronclient.v2_0.client [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.495618] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.495618] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.563518] env[65385]: DEBUG nova.policy [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '039fe0672f174065ba37b3f9c57e5d41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a3e5ed6b95f4282b1ef29d17fa58975', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 804.695770] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069478} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.696119] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.697042] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbf663a-658a-4dc6-866a-1ebe23df26a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.711941] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.712806] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.713632] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.730637] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 65f9bc55-7bab-43b0-a974-eb5080389b7d/65f9bc55-7bab-43b0-a974-eb5080389b7d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.731268] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c93a4ba-ced2-4f94-b1f9-5eef672a3864 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.760425] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c478db-93f1-2fe0-5467-ee036a5fc5b0, 'name': SearchDatastore_Task, 'duration_secs': 0.011149} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.762008] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 804.762763] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.762763] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.762763] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.762763] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.763224] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 804.763224] env[65385]: value = "task-4453676" [ 804.763224] env[65385]: _type = "Task" [ 804.763224] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.763505] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0845d6e5-8ca6-4f70-a017-f3ffa472bd9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.776581] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453676, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.781334] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.781647] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.782827] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f16472e9-7cff-425c-bbc1-7b1fb0285466 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.790296] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 804.790296] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b2ead-5072-0afb-8903-c1f253504283" [ 804.790296] env[65385]: _type = "Task" [ 804.790296] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.798063] env[65385]: DEBUG nova.compute.manager [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 804.802622] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0653fcdb-605a-4095-ad1f-38f54a98f058 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.805696] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b2ead-5072-0afb-8903-c1f253504283, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.836348] env[65385]: DEBUG oslo_vmware.api [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453675, 'name': PowerOffVM_Task, 'duration_secs': 0.483893} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.836640] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.836833] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.837264] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20cdbd52-5ae8-43ae-af51-090f804fe0b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.894072] env[65385]: DEBUG nova.network.neutron [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Successfully created port: fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 804.933280] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.934077] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.945554] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 804.945750] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 804.945926] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Deleting the datastore file [datastore2] e9fd0c9f-5795-44df-9dcb-982b7722ae69 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.946234] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-682f367f-059e-4a1a-be15-a8cbb99cd8e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.958530] env[65385]: DEBUG oslo_vmware.api [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453673, 'name': RemoveSnapshot_Task, 'duration_secs': 1.312018} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.960100] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 804.960249] env[65385]: INFO nova.compute.manager [None req-814e1055-e904-4e89-b6a4-e160c4020526 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Took 18.42 seconds to snapshot the instance on the hypervisor. [ 804.964084] env[65385]: DEBUG oslo_vmware.api [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for the task: (returnval){ [ 804.964084] env[65385]: value = "task-4453678" [ 804.964084] env[65385]: _type = "Task" [ 804.964084] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.975244] env[65385]: DEBUG oslo_vmware.api [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.980375] env[65385]: WARNING nova.compute.manager [None req-7dc51d1f-e598-478a-b782-d014554a2f7c tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Image not found during snapshot: nova.exception.ImageNotFound: Image 71310ad2-3ac2-4dcf-9d77-a75979646daf could not be found. [ 804.998109] env[65385]: DEBUG nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 805.076908] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.077600] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.077932] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.285091] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453676, 'name': ReconfigVM_Task, 'duration_secs': 0.452035} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.290260] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 65f9bc55-7bab-43b0-a974-eb5080389b7d/65f9bc55-7bab-43b0-a974-eb5080389b7d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.291734] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58c464c8-db0b-46ae-8bf2-1f5e71fba53f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.309803] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522b2ead-5072-0afb-8903-c1f253504283, 'name': SearchDatastore_Task, 'duration_secs': 0.011722} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.315279] env[65385]: DEBUG nova.network.neutron [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance_info_cache with network_info: [{"id": "ad95d76f-e023-47b2-a936-e7ed11779276", "address": "fa:16:3e:b0:c7:d2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95d76f-e0", "ovs_interfaceid": "ad95d76f-e023-47b2-a936-e7ed11779276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.320686] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 805.320686] env[65385]: value = "task-4453679" [ 805.320686] env[65385]: _type = "Task" [ 805.320686] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.323348] env[65385]: INFO nova.compute.manager [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] instance snapshotting [ 805.323348] env[65385]: WARNING nova.compute.manager [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 805.324364] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-854ce464-3f5e-4e0e-9eca-20af87b6abf7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.332657] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fda4cd-5ad7-4d95-8606-793adedbee1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.347175] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 805.347175] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5283d5fa-26a1-a37e-3bdb-b6f64ea9b43f" [ 805.347175] env[65385]: _type = "Task" [ 805.347175] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.369884] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453679, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.377354] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef50bfc2-865a-4c59-93f6-f7f565fd9e75 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.389368] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5283d5fa-26a1-a37e-3bdb-b6f64ea9b43f, 'name': SearchDatastore_Task, 'duration_secs': 0.013052} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.392300] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.392932] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 26a2d464-0aab-463f-8265-948a4f0ff188/26a2d464-0aab-463f-8265-948a4f0ff188.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.399285] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e8b3fa5-970d-470b-a6e3-de03b69cc370 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.409039] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 805.409039] env[65385]: value = "task-4453680" [ 805.409039] env[65385]: _type = "Task" [ 805.409039] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.420819] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453680, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.483343] env[65385]: DEBUG oslo_vmware.api [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Task: {'id': task-4453678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.438999} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.484810] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.485073] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 805.485292] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.485501] env[65385]: INFO nova.compute.manager [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Took 1.18 seconds to destroy the instance on the hypervisor. [ 805.485817] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 805.486396] env[65385]: DEBUG nova.compute.manager [-] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 805.487423] env[65385]: DEBUG nova.network.neutron [-] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 805.487782] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.488389] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.490364] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.541819] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.672649] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432b3847-c1ff-4cea-b6b9-539d00c67d2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.683307] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142a6ce8-710b-4f8c-977b-dd1d5a77566f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.717645] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1204a0a9-111a-4442-90b7-78b53be817be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.726777] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b905934-dedd-435a-93d9-548842a73e6b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.750746] env[65385]: DEBUG nova.compute.provider_tree [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.826651] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.844625] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453679, 'name': Rename_Task, 'duration_secs': 0.168915} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.844625] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.844625] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd62702b-2b52-4685-95ee-9bd33997c486 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.855511] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 805.855511] env[65385]: value = "task-4453681" [ 805.855511] env[65385]: _type = "Task" [ 805.855511] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.866327] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453681, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.904690] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 805.905062] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c71bb86f-a903-4baa-8ef2-c15ee8810731 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.915773] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 805.915773] env[65385]: value = "task-4453682" [ 805.915773] env[65385]: _type = "Task" [ 805.915773] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.924642] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453680, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508463} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.925556] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 26a2d464-0aab-463f-8265-948a4f0ff188/26a2d464-0aab-463f-8265-948a4f0ff188.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 805.926065] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.926065] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1450edb7-fb86-412b-b916-bf7f4dd4c2dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.932056] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453682, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.939023] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 805.939023] env[65385]: value = "task-4453683" [ 805.939023] env[65385]: _type = "Task" [ 805.939023] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.950734] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.999664] env[65385]: DEBUG nova.compute.manager [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Received event network-changed-622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 805.999953] env[65385]: DEBUG nova.compute.manager [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Refreshing instance network info cache due to event network-changed-622e8c02-7c83-4b27-9884-f6c399712a84. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 806.000954] env[65385]: DEBUG oslo_concurrency.lockutils [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Acquiring lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.000954] env[65385]: DEBUG oslo_concurrency.lockutils [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Acquired lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.000954] env[65385]: DEBUG nova.network.neutron [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Refreshing network info cache for port 622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 806.008915] env[65385]: DEBUG nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 806.038568] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 806.039035] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 806.039167] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 806.039368] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 806.039586] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 806.039768] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 806.039991] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.040180] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 806.040366] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 806.040618] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 806.040791] env[65385]: DEBUG nova.virt.hardware [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 806.042339] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a28271d-9bbc-4f8b-8ed3-1f121b86a847 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.053491] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4311db53-bb10-4e3e-aef8-2070d8233fa3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.144412] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.144700] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.257363] env[65385]: DEBUG nova.scheduler.client.report [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 806.336769] env[65385]: DEBUG nova.compute.manager [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65385) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:925}} [ 806.336769] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.370101] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.427552] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453682, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.450184] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07411} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.450445] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.451571] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890231da-aa07-4fe5-ba0d-202cacc9c289 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.457127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "ab8df1d8-615d-4083-bdf4-25f679a64982" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.457629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.458483] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "ab8df1d8-615d-4083-bdf4-25f679a64982-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.458950] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.459304] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.461655] env[65385]: INFO nova.compute.manager [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Terminating instance [ 806.486089] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 26a2d464-0aab-463f-8265-948a4f0ff188/26a2d464-0aab-463f-8265-948a4f0ff188.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.488272] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d185c88e-a2ab-4edc-99f5-3aeb7a68bbcb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.507415] env[65385]: WARNING neutronclient.v2_0.client [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 806.508438] env[65385]: WARNING openstack [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.510094] env[65385]: WARNING openstack [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.527223] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 806.527223] env[65385]: value = "task-4453684" [ 806.527223] env[65385]: _type = "Task" [ 806.527223] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.539126] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453684, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.601369] env[65385]: DEBUG nova.network.neutron [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Successfully updated port: fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 806.648980] env[65385]: DEBUG nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 806.656893] env[65385]: WARNING openstack [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.657419] env[65385]: WARNING openstack [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.665444] env[65385]: DEBUG nova.network.neutron [-] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 806.733442] env[65385]: WARNING neutronclient.v2_0.client [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 806.734874] env[65385]: WARNING openstack [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.735479] env[65385]: WARNING openstack [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.762907] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "98430d7e-626e-4532-802e-38f7fc583ae3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.763283] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "98430d7e-626e-4532-802e-38f7fc583ae3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.763504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "98430d7e-626e-4532-802e-38f7fc583ae3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.763719] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "98430d7e-626e-4532-802e-38f7fc583ae3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.763959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "98430d7e-626e-4532-802e-38f7fc583ae3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.766360] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.788s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.767096] env[65385]: DEBUG nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 806.772064] env[65385]: INFO nova.compute.manager [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Terminating instance [ 806.773263] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.151s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.774865] env[65385]: INFO nova.compute.claims [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.838992] env[65385]: DEBUG nova.network.neutron [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Updated VIF entry in instance network info cache for port 622e8c02-7c83-4b27-9884-f6c399712a84. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 806.839349] env[65385]: DEBUG nova.network.neutron [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Updating instance_info_cache with network_info: [{"id": "622e8c02-7c83-4b27-9884-f6c399712a84", "address": "fa:16:3e:a7:44:e2", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap622e8c02-7c", "ovs_interfaceid": "622e8c02-7c83-4b27-9884-f6c399712a84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 806.871943] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453681, 'name': PowerOnVM_Task} progress is 71%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.927871] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453682, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.989528] env[65385]: DEBUG nova.compute.manager [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 806.989643] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 806.990565] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec385951-8356-4641-94e1-26c75ee7a4d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.000188] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.000805] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4373f818-bc76-42f1-935a-ea657a0d4ad3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.009237] env[65385]: DEBUG oslo_vmware.api [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 807.009237] env[65385]: value = "task-4453685" [ 807.009237] env[65385]: _type = "Task" [ 807.009237] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.020109] env[65385]: DEBUG oslo_vmware.api [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.040200] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.106341] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.106519] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.106770] env[65385]: DEBUG nova.network.neutron [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 807.168960] env[65385]: INFO nova.compute.manager [-] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Took 1.68 seconds to deallocate network for instance. [ 807.178807] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.187625] env[65385]: DEBUG nova.compute.manager [req-0a2a34e0-5df1-4df5-9817-402d574d7a0c req-6944cd3e-ac94-48be-9ed6-b0a700bc8abc service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Received event network-vif-plugged-fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 807.187884] env[65385]: DEBUG oslo_concurrency.lockutils [req-0a2a34e0-5df1-4df5-9817-402d574d7a0c req-6944cd3e-ac94-48be-9ed6-b0a700bc8abc service nova] Acquiring lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.188154] env[65385]: DEBUG oslo_concurrency.lockutils [req-0a2a34e0-5df1-4df5-9817-402d574d7a0c req-6944cd3e-ac94-48be-9ed6-b0a700bc8abc service nova] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 807.188345] env[65385]: DEBUG oslo_concurrency.lockutils [req-0a2a34e0-5df1-4df5-9817-402d574d7a0c req-6944cd3e-ac94-48be-9ed6-b0a700bc8abc service nova] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 807.188562] env[65385]: DEBUG nova.compute.manager [req-0a2a34e0-5df1-4df5-9817-402d574d7a0c req-6944cd3e-ac94-48be-9ed6-b0a700bc8abc service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] No waiting events found dispatching network-vif-plugged-fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 807.188771] env[65385]: WARNING nova.compute.manager [req-0a2a34e0-5df1-4df5-9817-402d574d7a0c req-6944cd3e-ac94-48be-9ed6-b0a700bc8abc service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Received unexpected event network-vif-plugged-fe815bb4-491a-42f5-a5d3-2eba8997444a for instance with vm_state building and task_state spawning. [ 807.280157] env[65385]: DEBUG nova.compute.utils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 807.284663] env[65385]: DEBUG nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 807.286114] env[65385]: DEBUG nova.compute.manager [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 807.286439] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.287848] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d7b17e-a711-4ccd-a2e2-39b72fdfb1ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.298185] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.298494] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-175ea9ac-401b-4ddd-8a70-ce5350f309d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.307236] env[65385]: DEBUG oslo_vmware.api [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 807.307236] env[65385]: value = "task-4453686" [ 807.307236] env[65385]: _type = "Task" [ 807.307236] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.319108] env[65385]: DEBUG oslo_vmware.api [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453686, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.341950] env[65385]: DEBUG oslo_concurrency.lockutils [req-668c4cca-c8dd-4af3-ab1b-db636249eba5 req-df1bee7a-d78b-4c84-9976-9d64bc55f296 service nova] Releasing lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.373989] env[65385]: DEBUG oslo_vmware.api [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453681, 'name': PowerOnVM_Task, 'duration_secs': 1.285775} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.374348] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.374546] env[65385]: INFO nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Took 10.45 seconds to spawn the instance on the hypervisor. [ 807.374788] env[65385]: DEBUG nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 807.376246] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee2cc70-6101-4d91-abeb-c155d6391a4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.432102] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453682, 'name': CreateSnapshot_Task, 'duration_secs': 1.325929} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.432425] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 807.433469] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc7acc4-bbe3-4259-8d8f-6d47daf449ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.526079] env[65385]: DEBUG oslo_vmware.api [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453685, 'name': PowerOffVM_Task, 'duration_secs': 0.195974} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.527241] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 807.527241] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 807.527241] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75c6d7b9-7c87-40ea-9a38-72c4ae8ce153 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.540786] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453684, 'name': ReconfigVM_Task, 'duration_secs': 0.629666} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.540786] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 26a2d464-0aab-463f-8265-948a4f0ff188/26a2d464-0aab-463f-8265-948a4f0ff188.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.541291] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16ec1658-e980-4191-a1e7-da1fb6e02b42 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.549635] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 807.549635] env[65385]: value = "task-4453688" [ 807.549635] env[65385]: _type = "Task" [ 807.549635] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.563334] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453688, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.613159] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.613690] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.626911] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 807.626911] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 807.627186] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Deleting the datastore file [datastore2] ab8df1d8-615d-4083-bdf4-25f679a64982 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.627426] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d494d3ea-f35c-4680-9af0-3007d964e4e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.636107] env[65385]: DEBUG oslo_vmware.api [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for the task: (returnval){ [ 807.636107] env[65385]: value = "task-4453689" [ 807.636107] env[65385]: _type = "Task" [ 807.636107] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.646358] env[65385]: DEBUG oslo_vmware.api [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453689, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.661112] env[65385]: DEBUG nova.network.neutron [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 807.678584] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.687133] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.687687] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.762771] env[65385]: WARNING neutronclient.v2_0.client [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.763484] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.763982] env[65385]: WARNING openstack [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.787125] env[65385]: DEBUG nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 807.823368] env[65385]: DEBUG oslo_vmware.api [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453686, 'name': PowerOffVM_Task, 'duration_secs': 0.24664} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.826442] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 807.826637] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 807.827306] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f26ae90f-4c66-4b61-b9c6-c9e7257ce02b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.862954] env[65385]: DEBUG nova.network.neutron [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updating instance_info_cache with network_info: [{"id": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "address": "fa:16:3e:68:9f:df", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe815bb4-49", "ovs_interfaceid": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 807.898989] env[65385]: INFO nova.compute.manager [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Took 42.75 seconds to build instance. [ 807.949646] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 807.950070] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 807.950354] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleting the datastore file [datastore2] 98430d7e-626e-4532-802e-38f7fc583ae3 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.950799] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5fe6ee5b-3446-4d40-89c2-5e200e73d2f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.965469] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 807.971144] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f8eb3735-8d3d-44e0-9b34-7c0027ea5ab5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.985042] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 807.985042] env[65385]: value = "task-4453692" [ 807.985042] env[65385]: _type = "Task" [ 807.985042] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.985503] env[65385]: DEBUG oslo_vmware.api [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 807.985503] env[65385]: value = "task-4453691" [ 807.985503] env[65385]: _type = "Task" [ 807.985503] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.006777] env[65385]: DEBUG oslo_vmware.api [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453691, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.010680] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453692, 'name': CloneVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.064261] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453688, 'name': Rename_Task, 'duration_secs': 0.18557} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.064261] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 808.064261] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1dee3917-6052-4c96-80b3-acb869577c3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.074234] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 808.074234] env[65385]: value = "task-4453693" [ 808.074234] env[65385]: _type = "Task" [ 808.074234] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.090376] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.151102] env[65385]: DEBUG oslo_vmware.api [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Task: {'id': task-4453689, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178962} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.154262] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.154416] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.154620] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.154806] env[65385]: INFO nova.compute.manager [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Took 1.17 seconds to destroy the instance on the hypervisor. [ 808.155253] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 808.155704] env[65385]: DEBUG nova.compute.manager [-] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 808.155704] env[65385]: DEBUG nova.network.neutron [-] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 808.156128] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.156888] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.156888] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.256943] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.366054] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Releasing lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.366698] env[65385]: DEBUG nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance network_info: |[{"id": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "address": "fa:16:3e:68:9f:df", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe815bb4-49", "ovs_interfaceid": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 808.366967] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:9f:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '136c3499-9ca0-4f85-903d-1f194aa66ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe815bb4-491a-42f5-a5d3-2eba8997444a', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.379986] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 808.380898] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.381274] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6140c62f-dce7-4545-a7ce-bd36edb7a67f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.409562] env[65385]: DEBUG oslo_concurrency.lockutils [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 808.410095] env[65385]: DEBUG oslo_concurrency.lockutils [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.410224] env[65385]: DEBUG nova.compute.manager [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 808.410764] env[65385]: DEBUG oslo_concurrency.lockutils [None req-93dfb34e-b667-4c14-9ed5-b34150560e9d tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "65f9bc55-7bab-43b0-a974-eb5080389b7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.561s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.412482] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b52861d-2136-4ac0-8b1c-661358e5be05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.424314] env[65385]: DEBUG nova.compute.manager [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 808.425303] env[65385]: DEBUG nova.objects.instance [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'flavor' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.435539] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.435539] env[65385]: value = "task-4453694" [ 808.435539] env[65385]: _type = "Task" [ 808.435539] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.448014] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453694, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.501699] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41e84e1-0417-43d9-8c07-37e5a620fb16 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.514195] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453692, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.514576] env[65385]: DEBUG oslo_vmware.api [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453691, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198609} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.515316] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.515507] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.515761] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.516000] env[65385]: INFO nova.compute.manager [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Took 1.23 seconds to destroy the instance on the hypervisor. [ 808.516296] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 808.516514] env[65385]: DEBUG nova.compute.manager [-] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 808.516623] env[65385]: DEBUG nova.network.neutron [-] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 808.516899] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.517559] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.517835] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.529346] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e247fbd3-901f-42fa-8ecc-a8c2f46ee90a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.572734] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99f91ed-7c22-46aa-b3f1-ced15032f02d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.580073] env[65385]: DEBUG nova.compute.manager [req-3a5e6cb3-6376-4b3c-9dcc-873c57ec5a66 req-c1b247d9-64a4-4b50-97fc-be2c3866ac85 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received event network-vif-deleted-6f259efd-d33d-4c33-8f8d-a98afc1ec5b8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 808.580362] env[65385]: DEBUG nova.compute.manager [req-3a5e6cb3-6376-4b3c-9dcc-873c57ec5a66 req-c1b247d9-64a4-4b50-97fc-be2c3866ac85 service nova] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Received event network-vif-deleted-7217d12a-b81f-4236-971b-a59c2709cf6f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 808.598240] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453693, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.601969] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a9b1b8-9f78-43e4-bba4-a66e664e0c27 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.625079] env[65385]: DEBUG nova.compute.provider_tree [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.700525] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.799536] env[65385]: DEBUG nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 808.840707] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 808.841015] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 808.841223] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 808.841355] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 808.841541] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 808.841664] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 808.841817] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.842020] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 808.842258] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 808.842813] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 808.842813] env[65385]: DEBUG nova.virt.hardware [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 808.843676] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88482b93-397c-4812-bee6-0200d962b263 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.860138] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600372d5-30e3-4772-be14-0b2b4d00c63d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.880252] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.885334] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 808.885751] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.886013] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-86e6e29c-3b9e-4d62-b806-38f1db94d4cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.906367] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.906367] env[65385]: value = "task-4453695" [ 808.906367] env[65385]: _type = "Task" [ 808.906367] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.916149] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453695, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.948477] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453694, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.009613] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453692, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.091731] env[65385]: DEBUG oslo_vmware.api [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453693, 'name': PowerOnVM_Task, 'duration_secs': 0.655186} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.092256] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 809.092256] env[65385]: INFO nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Took 8.51 seconds to spawn the instance on the hypervisor. [ 809.092376] env[65385]: DEBUG nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 809.093282] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69baea91-b9e8-417a-b676-610471babb29 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.113499] env[65385]: DEBUG nova.network.neutron [-] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 809.129550] env[65385]: DEBUG nova.scheduler.client.report [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 809.418601] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453695, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.443676] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.444018] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ead657ee-4b8c-4944-abd0-a5dba6ef411f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.453876] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453694, 'name': CreateVM_Task, 'duration_secs': 0.690001} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.455937] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.456401] env[65385]: DEBUG oslo_vmware.api [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 809.456401] env[65385]: value = "task-4453696" [ 809.456401] env[65385]: _type = "Task" [ 809.456401] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.457063] env[65385]: WARNING neutronclient.v2_0.client [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.457476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.457649] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.458042] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 809.458353] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50218043-1ae7-4529-a33e-c1e716d40ed1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.475729] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 809.475729] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529b789a-ca61-5935-7cb9-6ac867ee3573" [ 809.475729] env[65385]: _type = "Task" [ 809.475729] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.476223] env[65385]: DEBUG oslo_vmware.api [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453696, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.489351] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529b789a-ca61-5935-7cb9-6ac867ee3573, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.501252] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453692, 'name': CloneVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.502978] env[65385]: DEBUG nova.network.neutron [-] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 809.542475] env[65385]: DEBUG nova.compute.manager [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Received event network-changed-fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 809.542680] env[65385]: DEBUG nova.compute.manager [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Refreshing instance network info cache due to event network-changed-fe815bb4-491a-42f5-a5d3-2eba8997444a. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 809.542895] env[65385]: DEBUG oslo_concurrency.lockutils [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Acquiring lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.543041] env[65385]: DEBUG oslo_concurrency.lockutils [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Acquired lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.543198] env[65385]: DEBUG nova.network.neutron [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Refreshing network info cache for port fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 809.616157] env[65385]: INFO nova.compute.manager [-] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Took 1.46 seconds to deallocate network for instance. [ 809.617075] env[65385]: INFO nova.compute.manager [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Took 42.51 seconds to build instance. [ 809.637226] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.864s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.638864] env[65385]: DEBUG nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 809.641047] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.906s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.642767] env[65385]: INFO nova.compute.claims [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.918947] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453695, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.969750] env[65385]: DEBUG oslo_vmware.api [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453696, 'name': PowerOffVM_Task, 'duration_secs': 0.351596} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.969957] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.970180] env[65385]: DEBUG nova.compute.manager [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 809.970953] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9097f70c-6269-4d1b-85ec-b1c276b72a3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.989692] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529b789a-ca61-5935-7cb9-6ac867ee3573, 'name': SearchDatastore_Task, 'duration_secs': 0.017583} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.989887] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.990073] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.990620] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.990842] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.991017] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.991364] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bdbd6401-843e-4765-9882-3b7898819fb4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.005501] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453692, 'name': CloneVM_Task, 'duration_secs': 1.615678} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.005501] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Created linked-clone VM from snapshot [ 810.006145] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822c5a03-bb3b-4145-a223-02a73ab8cdd0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.010983] env[65385]: INFO nova.compute.manager [-] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Took 1.49 seconds to deallocate network for instance. [ 810.011342] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.011565] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.014983] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e22ef8bd-3a9d-4a2d-8dc4-7d565a3280cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.025637] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Uploading image 5e8408d4-fb50-4b8f-95d6-5374d829517e {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 810.035284] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 810.035284] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520262fd-d656-4de0-5308-0fc1de70f71a" [ 810.035284] env[65385]: _type = "Task" [ 810.035284] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.045855] env[65385]: WARNING neutronclient.v2_0.client [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.046704] env[65385]: WARNING openstack [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.047087] env[65385]: WARNING openstack [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.054840] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520262fd-d656-4de0-5308-0fc1de70f71a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.063719] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 810.063719] env[65385]: value = "vm-871046" [ 810.063719] env[65385]: _type = "VirtualMachine" [ 810.063719] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 810.064654] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3837adbb-959b-49b7-a257-457dd4d65fe0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.074217] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease: (returnval){ [ 810.074217] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f49dbc-7893-07f0-a584-bfadee96d353" [ 810.074217] env[65385]: _type = "HttpNfcLease" [ 810.074217] env[65385]: } obtained for exporting VM: (result){ [ 810.074217] env[65385]: value = "vm-871046" [ 810.074217] env[65385]: _type = "VirtualMachine" [ 810.074217] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 810.074543] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the lease: (returnval){ [ 810.074543] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f49dbc-7893-07f0-a584-bfadee96d353" [ 810.074543] env[65385]: _type = "HttpNfcLease" [ 810.074543] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 810.083615] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 810.083615] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f49dbc-7893-07f0-a584-bfadee96d353" [ 810.083615] env[65385]: _type = "HttpNfcLease" [ 810.083615] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 810.120405] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0c0a456b-fa05-4139-ab50-61bbcc22e4aa tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.670s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.130154] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.148865] env[65385]: DEBUG nova.compute.utils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 810.157019] env[65385]: DEBUG nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 810.157244] env[65385]: DEBUG nova.network.neutron [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 810.157619] env[65385]: WARNING neutronclient.v2_0.client [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.157939] env[65385]: WARNING neutronclient.v2_0.client [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.159266] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.159266] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.185370] env[65385]: WARNING openstack [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.185840] env[65385]: WARNING openstack [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.222020] env[65385]: DEBUG nova.policy [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b26caf958b047c79ed5d7f3d4d6b883', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '647f4e3157c34a8f9a64cb8ba8b72da6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 810.277722] env[65385]: WARNING neutronclient.v2_0.client [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.278632] env[65385]: WARNING openstack [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.279124] env[65385]: WARNING openstack [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.317148] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "0115dde8-b0aa-4960-a452-ced6587bc567" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.317427] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "0115dde8-b0aa-4960-a452-ced6587bc567" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.318198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "0115dde8-b0aa-4960-a452-ced6587bc567-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.318198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "0115dde8-b0aa-4960-a452-ced6587bc567-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.318198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "0115dde8-b0aa-4960-a452-ced6587bc567-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.320589] env[65385]: INFO nova.compute.manager [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Terminating instance [ 810.406084] env[65385]: DEBUG nova.network.neutron [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updated VIF entry in instance network info cache for port fe815bb4-491a-42f5-a5d3-2eba8997444a. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 810.406084] env[65385]: DEBUG nova.network.neutron [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updating instance_info_cache with network_info: [{"id": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "address": "fa:16:3e:68:9f:df", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe815bb4-49", "ovs_interfaceid": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 810.421305] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453695, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.486665] env[65385]: DEBUG oslo_concurrency.lockutils [None req-894d2636-6fa2-4a3f-b3e5-fbc64792caf9 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.077s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.532936] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.547650] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520262fd-d656-4de0-5308-0fc1de70f71a, 'name': SearchDatastore_Task, 'duration_secs': 0.016709} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.548939] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b82d91d8-260a-4189-8664-8d498c11693c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.555929] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 810.555929] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529da75b-b6a2-9a22-c65e-5ccc4babcdd9" [ 810.555929] env[65385]: _type = "Task" [ 810.555929] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.565696] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529da75b-b6a2-9a22-c65e-5ccc4babcdd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.585271] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 810.585271] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f49dbc-7893-07f0-a584-bfadee96d353" [ 810.585271] env[65385]: _type = "HttpNfcLease" [ 810.585271] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 810.585271] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 810.585271] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f49dbc-7893-07f0-a584-bfadee96d353" [ 810.585271] env[65385]: _type = "HttpNfcLease" [ 810.585271] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 810.586124] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdb54b8-c323-4419-bc49-34408d87e297 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.595886] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521ac185-c5d8-cc40-be0d-75322554c767/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 810.596230] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521ac185-c5d8-cc40-be0d-75322554c767/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 810.663891] env[65385]: DEBUG nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 810.667387] env[65385]: DEBUG nova.network.neutron [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Successfully created port: bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 810.746805] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d129660f-47f5-4654-a146-410b1e728656 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.830268] env[65385]: DEBUG nova.compute.manager [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 810.830474] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.831905] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6970a90-6042-4959-ae8a-29d4a193e2e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.840304] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.842989] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4306c15-0476-474e-be5e-8270041b7b2e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.850573] env[65385]: DEBUG oslo_vmware.api [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 810.850573] env[65385]: value = "task-4453698" [ 810.850573] env[65385]: _type = "Task" [ 810.850573] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.862476] env[65385]: DEBUG oslo_vmware.api [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.909767] env[65385]: DEBUG oslo_concurrency.lockutils [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] Releasing lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.909982] env[65385]: DEBUG nova.compute.manager [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Received event network-vif-deleted-3ed3260d-d4bf-49a0-baa3-bb749653df5e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 810.910195] env[65385]: DEBUG nova.compute.manager [req-2281e6e3-bf76-449f-b0ef-f497cd5f2aa9 req-88154536-d7ea-4394-8ff4-4537091f2cd0 service nova] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Received event network-vif-deleted-a150e5dc-1294-45e5-9346-d31c443b7efe {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 810.923489] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453695, 'name': CreateVM_Task, 'duration_secs': 1.539642} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.926520] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.927458] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.927654] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.927985] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 810.928271] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb9ba3cf-cc55-4378-876b-9bb9803eccae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.935638] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 810.935638] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5206d1e7-361d-82dc-a196-eca7450ce066" [ 810.935638] env[65385]: _type = "Task" [ 810.935638] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.950264] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5206d1e7-361d-82dc-a196-eca7450ce066, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.070749] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529da75b-b6a2-9a22-c65e-5ccc4babcdd9, 'name': SearchDatastore_Task, 'duration_secs': 0.018443} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.071029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.071295] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] ad29afd9-3e52-4fac-977f-91774acad7d0/ad29afd9-3e52-4fac-977f-91774acad7d0.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.071567] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a55fb58-645a-4cdb-b7c2-c89fbf6da365 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.081142] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 811.081142] env[65385]: value = "task-4453699" [ 811.081142] env[65385]: _type = "Task" [ 811.081142] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.102360] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.316624] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789c82e0-25fb-4292-9b40-398f121d5558 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.326703] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec68663-4974-43f5-89a6-9067b45a405e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.388571] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec28c8c-437b-427c-b837-dcb08a410ce9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.403090] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403514ea-22de-4e29-8506-be5503f522ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.407841] env[65385]: DEBUG oslo_vmware.api [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453698, 'name': PowerOffVM_Task, 'duration_secs': 0.318408} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.408266] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 811.408443] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 811.409552] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d01bdc4-ccf0-4e9f-9ea9-dedee1091557 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.432303] env[65385]: DEBUG nova.compute.provider_tree [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.451260] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5206d1e7-361d-82dc-a196-eca7450ce066, 'name': SearchDatastore_Task, 'duration_secs': 0.015207} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.455027] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.455027] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.455027] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.455027] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.455027] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.455401] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6e44330-618b-4a78-9f71-4aa0222347f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.476477] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.476662] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.477667] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f79efb7f-5bb3-43e9-8725-812b3391493c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.487892] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 811.487892] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e05dc5-a383-ad2c-69d2-5be38d5ee40b" [ 811.487892] env[65385]: _type = "Task" [ 811.487892] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.497221] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 811.497641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 811.497874] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Deleting the datastore file [datastore2] 0115dde8-b0aa-4960-a452-ced6587bc567 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 811.498430] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59eeca92-86a0-45fa-87c2-377e2ad42ed4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.505071] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e05dc5-a383-ad2c-69d2-5be38d5ee40b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.514470] env[65385]: DEBUG oslo_vmware.api [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for the task: (returnval){ [ 811.514470] env[65385]: value = "task-4453701" [ 811.514470] env[65385]: _type = "Task" [ 811.514470] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.526385] env[65385]: DEBUG oslo_vmware.api [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.592975] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453699, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.687371] env[65385]: DEBUG nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 811.716864] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 811.721562] env[65385]: DEBUG nova.virt.hardware [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 811.722045] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc4f2e9-a33f-4ce4-88ed-8f44eee7df0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.733136] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a010977b-34f7-4f24-a10b-56ae82d863ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.825043] env[65385]: DEBUG nova.compute.manager [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Received event network-changed-622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 811.825136] env[65385]: DEBUG nova.compute.manager [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Refreshing instance network info cache due to event network-changed-622e8c02-7c83-4b27-9884-f6c399712a84. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 811.825746] env[65385]: DEBUG oslo_concurrency.lockutils [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Acquiring lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.825746] env[65385]: DEBUG oslo_concurrency.lockutils [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Acquired lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.826219] env[65385]: DEBUG nova.network.neutron [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Refreshing network info cache for port 622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 811.937353] env[65385]: DEBUG nova.scheduler.client.report [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 812.000876] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e05dc5-a383-ad2c-69d2-5be38d5ee40b, 'name': SearchDatastore_Task, 'duration_secs': 0.050813} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.001716] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b642897c-5eb0-42b6-9bfa-0dcdb2ead026 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.015028] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 812.015028] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bac120-c2fe-8dc0-05d2-b406dd0959d0" [ 812.015028] env[65385]: _type = "Task" [ 812.015028] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.034337] env[65385]: DEBUG oslo_vmware.api [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Task: {'id': task-4453701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315782} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.034613] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bac120-c2fe-8dc0-05d2-b406dd0959d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.034887] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 812.035109] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 812.035379] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 812.035601] env[65385]: INFO nova.compute.manager [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Took 1.21 seconds to destroy the instance on the hypervisor. [ 812.035878] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 812.036450] env[65385]: DEBUG nova.compute.manager [-] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 812.037274] env[65385]: DEBUG nova.network.neutron [-] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 812.037596] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.039023] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.039023] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.096054] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453699, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608926} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.096054] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] ad29afd9-3e52-4fac-977f-91774acad7d0/ad29afd9-3e52-4fac-977f-91774acad7d0.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.096054] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.096054] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4af8cffb-f0e4-4477-a275-89ae23ec53b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.108473] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 812.108473] env[65385]: value = "task-4453702" [ 812.108473] env[65385]: _type = "Task" [ 812.108473] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.118310] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.126312] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.326975] env[65385]: DEBUG nova.network.neutron [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Successfully updated port: bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 812.330566] env[65385]: WARNING neutronclient.v2_0.client [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.331318] env[65385]: WARNING openstack [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.331662] env[65385]: WARNING openstack [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.416262] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "59450963-d83f-46e8-8c13-05d4f1818c64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.416601] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "59450963-d83f-46e8-8c13-05d4f1818c64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.416849] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "59450963-d83f-46e8-8c13-05d4f1818c64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.417070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "59450963-d83f-46e8-8c13-05d4f1818c64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.417388] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "59450963-d83f-46e8-8c13-05d4f1818c64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.420988] env[65385]: INFO nova.compute.manager [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Terminating instance [ 812.444903] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.804s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.446129] env[65385]: DEBUG nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 812.450553] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.684s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.450836] env[65385]: DEBUG nova.objects.instance [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lazy-loading 'resources' on Instance uuid 4232f3a4-c42b-4b9c-8e1d-da591634ffbb {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.523536] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bac120-c2fe-8dc0-05d2-b406dd0959d0, 'name': SearchDatastore_Task, 'duration_secs': 0.019787} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.524463] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 812.524742] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 812.525082] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e7bbe03-9d9f-43b8-9218-67358e44750a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.529431] env[65385]: WARNING openstack [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.529707] env[65385]: WARNING openstack [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.544090] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 812.544090] env[65385]: value = "task-4453703" [ 812.544090] env[65385]: _type = "Task" [ 812.544090] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.554700] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.622108] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075918} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.622108] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.623463] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e12239-ef68-4a7f-9015-8b54c4d3d5bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.627165] env[65385]: WARNING neutronclient.v2_0.client [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.628151] env[65385]: WARNING openstack [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.628616] env[65385]: WARNING openstack [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.659587] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] ad29afd9-3e52-4fac-977f-91774acad7d0/ad29afd9-3e52-4fac-977f-91774acad7d0.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.659838] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32ab58fc-e1ec-41f9-aa22-f62dc516b291 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.685804] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 812.685804] env[65385]: value = "task-4453704" [ 812.685804] env[65385]: _type = "Task" [ 812.685804] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.698360] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453704, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.720018] env[65385]: DEBUG nova.network.neutron [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Updated VIF entry in instance network info cache for port 622e8c02-7c83-4b27-9884-f6c399712a84. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 812.720468] env[65385]: DEBUG nova.network.neutron [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Updating instance_info_cache with network_info: [{"id": "622e8c02-7c83-4b27-9884-f6c399712a84", "address": "fa:16:3e:a7:44:e2", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap622e8c02-7c", "ovs_interfaceid": "622e8c02-7c83-4b27-9884-f6c399712a84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 812.829465] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.829694] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 812.829970] env[65385]: DEBUG nova.network.neutron [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 812.840311] env[65385]: DEBUG nova.compute.manager [req-688037b7-bdb3-4f1d-ad99-448017b80cc9 req-041da0d5-5df5-41dd-9b12-b83aa3af301c service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received event network-vif-plugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 812.840311] env[65385]: DEBUG oslo_concurrency.lockutils [req-688037b7-bdb3-4f1d-ad99-448017b80cc9 req-041da0d5-5df5-41dd-9b12-b83aa3af301c service nova] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.840758] env[65385]: DEBUG oslo_concurrency.lockutils [req-688037b7-bdb3-4f1d-ad99-448017b80cc9 req-041da0d5-5df5-41dd-9b12-b83aa3af301c service nova] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.840758] env[65385]: DEBUG oslo_concurrency.lockutils [req-688037b7-bdb3-4f1d-ad99-448017b80cc9 req-041da0d5-5df5-41dd-9b12-b83aa3af301c service nova] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.840905] env[65385]: DEBUG nova.compute.manager [req-688037b7-bdb3-4f1d-ad99-448017b80cc9 req-041da0d5-5df5-41dd-9b12-b83aa3af301c service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] No waiting events found dispatching network-vif-plugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 812.841022] env[65385]: WARNING nova.compute.manager [req-688037b7-bdb3-4f1d-ad99-448017b80cc9 req-041da0d5-5df5-41dd-9b12-b83aa3af301c service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received unexpected event network-vif-plugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 for instance with vm_state building and task_state spawning. [ 812.925859] env[65385]: DEBUG nova.compute.manager [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 812.926240] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 812.927322] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e48cb08-fe65-4187-af12-12021075bc2e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.937214] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.937841] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08064fad-aad8-475c-9577-ac92df4b8c45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.947037] env[65385]: DEBUG oslo_vmware.api [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 812.947037] env[65385]: value = "task-4453705" [ 812.947037] env[65385]: _type = "Task" [ 812.947037] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.955240] env[65385]: DEBUG nova.compute.utils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 812.961989] env[65385]: DEBUG nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 812.961989] env[65385]: DEBUG nova.network.neutron [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 812.961989] env[65385]: WARNING neutronclient.v2_0.client [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.962145] env[65385]: WARNING neutronclient.v2_0.client [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.962694] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.963133] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.979540] env[65385]: DEBUG oslo_vmware.api [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453705, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.984956] env[65385]: DEBUG nova.network.neutron [-] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 813.032045] env[65385]: DEBUG nova.policy [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16e02a4d719145cba9bce20e1c8ed3d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dde1f88f1730420580061f4068e645ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 813.063248] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453703, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.201341] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453704, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.226771] env[65385]: DEBUG oslo_concurrency.lockutils [req-66e8bffe-eded-4fdc-aca4-c1024285926e req-b068cd88-6ac5-4805-b07f-172a574ad413 service nova] Releasing lock "refresh_cache-26a2d464-0aab-463f-8265-948a4f0ff188" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 813.335869] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.336239] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.437875] env[65385]: DEBUG nova.network.neutron [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 813.462678] env[65385]: DEBUG nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 813.465676] env[65385]: DEBUG oslo_vmware.api [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453705, 'name': PowerOffVM_Task, 'duration_secs': 0.463735} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.469179] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 813.469385] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 813.470411] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6afb6b4-3cb5-4e17-904a-23a899414bdf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.486203] env[65385]: INFO nova.compute.manager [-] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Took 1.45 seconds to deallocate network for instance. [ 813.539538] env[65385]: DEBUG nova.network.neutron [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Successfully created port: e615249d-2b37-4211-abe1-70015ea3d4cf {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 813.560730] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.768496} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.564486] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 813.566163] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.569242] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.569372] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.581533] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d501600c-285b-48f9-9e1b-ef454be16e90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.583977] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 813.584181] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 813.584359] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Deleting the datastore file [datastore2] 59450963-d83f-46e8-8c13-05d4f1818c64 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.588298] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-546a7b00-aa72-4ddd-9497-342d54655b34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.599412] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 813.599412] env[65385]: value = "task-4453707" [ 813.599412] env[65385]: _type = "Task" [ 813.599412] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.601202] env[65385]: DEBUG oslo_vmware.api [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for the task: (returnval){ [ 813.601202] env[65385]: value = "task-4453708" [ 813.601202] env[65385]: _type = "Task" [ 813.601202] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.620984] env[65385]: DEBUG oslo_vmware.api [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453708, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.621171] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453707, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.667991] env[65385]: WARNING neutronclient.v2_0.client [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 813.668129] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.668591] env[65385]: WARNING openstack [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.701184] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453704, 'name': ReconfigVM_Task, 'duration_secs': 0.540761} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.701529] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Reconfigured VM instance instance-00000038 to attach disk [datastore1] ad29afd9-3e52-4fac-977f-91774acad7d0/ad29afd9-3e52-4fac-977f-91774acad7d0.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.702283] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b27d365f-02ec-4a75-9c9a-0be4f4f02a30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.711153] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 813.711153] env[65385]: value = "task-4453709" [ 813.711153] env[65385]: _type = "Task" [ 813.711153] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.713055] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281f9fce-25af-4f38-aa6c-4dab44dc9195 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.733932] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab590549-9a66-4c2f-9e20-956223691156 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.740801] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453709, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.774913] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1d9d9a-1037-4710-88fa-170b0d257ab4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.782489] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdcd7f8-c301-43b9-b9ea-405f6eb2f6d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.789291] env[65385]: DEBUG nova.network.neutron [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbadc19c-78", "ovs_interfaceid": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 813.803015] env[65385]: DEBUG nova.compute.provider_tree [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.907644] env[65385]: DEBUG nova.compute.manager [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Stashing vm_state: stopped {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 813.996320] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.115416] env[65385]: DEBUG oslo_vmware.api [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Task: {'id': task-4453708, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.418295} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.118748] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.118899] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.119119] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.119300] env[65385]: INFO nova.compute.manager [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Took 1.19 seconds to destroy the instance on the hypervisor. [ 814.119533] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 814.119739] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453707, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077344} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.119937] env[65385]: DEBUG nova.compute.manager [-] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 814.120049] env[65385]: DEBUG nova.network.neutron [-] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 814.120343] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.120892] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.121157] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.127770] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.128890] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c05cb48-7c26-4484-89eb-45b0dd7e01a0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.153269] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.153909] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7736cb1-d271-4cde-a23a-5075f4f24e99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.174892] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 814.174892] env[65385]: value = "task-4453710" [ 814.174892] env[65385]: _type = "Task" [ 814.174892] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.185731] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453710, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.187925] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.222977] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453709, 'name': Rename_Task, 'duration_secs': 0.286404} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.223345] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 814.223521] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8487a30e-1cf0-475e-88a9-87564fbeacb4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.235781] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 814.235781] env[65385]: value = "task-4453711" [ 814.235781] env[65385]: _type = "Task" [ 814.235781] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.246569] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.295124] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 814.295480] env[65385]: DEBUG nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Instance network_info: |[{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbadc19c-78", "ovs_interfaceid": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 814.295978] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:ff:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4734e5e-2a76-4bda-8905-70c9bf9e007f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bbadc19c-78ee-4246-a53d-77f01c5ccea8', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.303971] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating folder: Project (647f4e3157c34a8f9a64cb8ba8b72da6). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.304354] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be9f8056-b2d8-47d6-8553-81c389e54412 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.307494] env[65385]: DEBUG nova.scheduler.client.report [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 814.322715] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Created folder: Project (647f4e3157c34a8f9a64cb8ba8b72da6) in parent group-v870881. [ 814.322801] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating folder: Instances. Parent ref: group-v871049. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 814.324141] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-663a0d5d-f053-469a-9673-e33c56271298 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.334950] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Created folder: Instances in parent group-v871049. [ 814.335234] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 814.335440] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 814.335688] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee3198d4-cfd4-43d4-863c-efd0fca84c4d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.365345] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.365345] env[65385]: value = "task-4453714" [ 814.365345] env[65385]: _type = "Task" [ 814.365345] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.376403] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453714, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.433676] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.475843] env[65385]: DEBUG nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 814.502338] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 814.502610] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 814.502804] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 814.503114] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 814.503325] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 814.503484] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 814.503973] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.504185] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 814.504361] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 814.504575] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 814.504764] env[65385]: DEBUG nova.virt.hardware [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 814.505738] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7157c69-4366-49af-b394-8b478d55556e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.515456] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425cd070-9f0a-45c0-a4cd-648248aeb7e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.686251] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453710, 'name': ReconfigVM_Task, 'duration_secs': 0.323205} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.686673] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Reconfigured VM instance instance-00000039 to attach disk [datastore1] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 814.687346] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ae60ac8-d3f5-4318-a28b-6f6c2c2867ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.695356] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 814.695356] env[65385]: value = "task-4453715" [ 814.695356] env[65385]: _type = "Task" [ 814.695356] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.710093] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453715, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.750551] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453711, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.814462] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.364s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.818297] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.611s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.820828] env[65385]: DEBUG nova.objects.instance [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lazy-loading 'resources' on Instance uuid 649b4e21-fc55-415c-a1f3-ec724397b874 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.843251] env[65385]: INFO nova.scheduler.client.report [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Deleted allocations for instance 4232f3a4-c42b-4b9c-8e1d-da591634ffbb [ 814.878471] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453714, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.909274] env[65385]: DEBUG nova.compute.manager [req-de0a7996-1345-4a50-9397-0c8eebecf075 req-f472937c-2d2f-44c0-9b2c-147d08a8843e service nova] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Received event network-vif-deleted-e37bf46a-308d-44a5-b573-64c3b7aa5a73 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 815.191621] env[65385]: DEBUG nova.network.neutron [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Successfully updated port: e615249d-2b37-4211-abe1-70015ea3d4cf {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 815.212979] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453715, 'name': Rename_Task, 'duration_secs': 0.157824} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.213823] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.213995] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f2ee289-fbf0-44ee-9a54-b200b4bc3d45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.229024] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 815.229024] env[65385]: value = "task-4453716" [ 815.229024] env[65385]: _type = "Task" [ 815.229024] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.241406] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.251724] env[65385]: DEBUG oslo_vmware.api [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453711, 'name': PowerOnVM_Task, 'duration_secs': 0.768208} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.252014] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.252217] env[65385]: INFO nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Took 9.24 seconds to spawn the instance on the hypervisor. [ 815.252397] env[65385]: DEBUG nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 815.253234] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba94c11-8921-4526-80d1-b99324e32e62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.275107] env[65385]: DEBUG nova.network.neutron [-] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 815.354911] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6c454857-08cd-4592-9900-4b7ad57318cc tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "4232f3a4-c42b-4b9c-8e1d-da591634ffbb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.824s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.382312] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453714, 'name': CreateVM_Task, 'duration_secs': 0.542979} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.385548] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.386360] env[65385]: WARNING neutronclient.v2_0.client [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.389119] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.389119] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 815.389119] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 815.389119] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8b6cae7-5c0f-4bf4-8fbe-aa94efb90444 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.393677] env[65385]: DEBUG nova.compute.manager [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received event network-changed-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 815.394366] env[65385]: DEBUG nova.compute.manager [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Refreshing instance network info cache due to event network-changed-bbadc19c-78ee-4246-a53d-77f01c5ccea8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 815.394366] env[65385]: DEBUG oslo_concurrency.lockutils [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Acquiring lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.394366] env[65385]: DEBUG oslo_concurrency.lockutils [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Acquired lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 815.394366] env[65385]: DEBUG nova.network.neutron [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Refreshing network info cache for port bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 815.398277] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 815.398277] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525af01a-6af5-9d62-eab5-d2c5ac59f592" [ 815.398277] env[65385]: _type = "Task" [ 815.398277] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.412083] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525af01a-6af5-9d62-eab5-d2c5ac59f592, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.618781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Acquiring lock "1ad1e02b-fb82-4a88-9047-be316d590bd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.619127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.619127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Acquiring lock "1ad1e02b-fb82-4a88-9047-be316d590bd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.619959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.619959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.621835] env[65385]: INFO nova.compute.manager [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Terminating instance [ 815.698668] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "refresh_cache-af8d01e2-fb86-41c4-99a5-204b30eeda0c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.698918] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquired lock "refresh_cache-af8d01e2-fb86-41c4-99a5-204b30eeda0c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 815.699189] env[65385]: DEBUG nova.network.neutron [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 815.714654] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "483cf486-f0fc-48a4-9db2-970d0f35865a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.714858] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.715052] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "483cf486-f0fc-48a4-9db2-970d0f35865a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.715241] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.715407] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.718084] env[65385]: INFO nova.compute.manager [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Terminating instance [ 815.738336] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453716, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.780450] env[65385]: INFO nova.compute.manager [-] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Took 1.66 seconds to deallocate network for instance. [ 815.782900] env[65385]: INFO nova.compute.manager [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Took 40.54 seconds to build instance. [ 815.889942] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad72253e-741c-4f8c-9d40-ad8709327f34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.900023] env[65385]: WARNING neutronclient.v2_0.client [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.900023] env[65385]: WARNING openstack [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.900023] env[65385]: WARNING openstack [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.912481] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adebcdc9-9946-4caf-af75-343dd06227eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.952995] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525af01a-6af5-9d62-eab5-d2c5ac59f592, 'name': SearchDatastore_Task, 'duration_secs': 0.014866} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.960310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 815.960310] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.960310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.960310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 815.960310] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.961263] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d761652b-d6cf-48db-9f32-5d8ca63835d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.964370] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b40bfe2a-0dd4-493a-9c91-3461754100d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.976087] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a10643-d1e6-45ca-a820-54f120caa67e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.983027] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.983027] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.983350] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d059f21-a033-40d7-a3e4-87b17c4525d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.994966] env[65385]: DEBUG nova.compute.provider_tree [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.001289] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 816.001289] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527048d2-e5ac-d333-15cc-b4b8cb22618a" [ 816.001289] env[65385]: _type = "Task" [ 816.001289] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.010694] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527048d2-e5ac-d333-15cc-b4b8cb22618a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.066386] env[65385]: WARNING openstack [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.066787] env[65385]: WARNING openstack [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.128028] env[65385]: DEBUG nova.compute.manager [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 816.128134] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.128504] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c880c74b-2bb3-4104-b355-5512014286de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.137486] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 816.137486] env[65385]: value = "task-4453717" [ 816.137486] env[65385]: _type = "Task" [ 816.137486] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.142634] env[65385]: WARNING neutronclient.v2_0.client [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.143452] env[65385]: WARNING openstack [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.144192] env[65385]: WARNING openstack [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.162341] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453717, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.203239] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.203822] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.227118] env[65385]: DEBUG nova.compute.manager [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 816.227422] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.229351] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f3470c-305b-42fb-b9ab-edf16187b666 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.243235] env[65385]: DEBUG oslo_vmware.api [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453716, 'name': PowerOnVM_Task, 'duration_secs': 0.51813} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.246718] env[65385]: DEBUG nova.network.neutron [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 816.249210] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.249452] env[65385]: INFO nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Took 7.45 seconds to spawn the instance on the hypervisor. [ 816.249658] env[65385]: DEBUG nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 816.250017] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.251124] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2340ebcf-a502-4635-836e-272c04361bb5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.254066] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9978ebd-022a-4bbb-a35e-0e671a394ba7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.266941] env[65385]: DEBUG oslo_vmware.api [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 816.266941] env[65385]: value = "task-4453718" [ 816.266941] env[65385]: _type = "Task" [ 816.266941] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.274072] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.274379] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.284409] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83d0b1a7-2ae2-458b-8250-d8418be82968 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.047s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.284845] env[65385]: DEBUG oslo_vmware.api [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.287871] env[65385]: DEBUG nova.network.neutron [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updated VIF entry in instance network info cache for port bbadc19c-78ee-4246-a53d-77f01c5ccea8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 816.288305] env[65385]: DEBUG nova.network.neutron [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbadc19c-78", "ovs_interfaceid": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 816.290390] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.348730] env[65385]: WARNING neutronclient.v2_0.client [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.349448] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.349827] env[65385]: WARNING openstack [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.464705] env[65385]: DEBUG nova.network.neutron [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Updating instance_info_cache with network_info: [{"id": "e615249d-2b37-4211-abe1-70015ea3d4cf", "address": "fa:16:3e:a4:42:90", "network": {"id": "109b5b1e-bc16-42c7-a4fe-3f5b85a52cfa", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-704603508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dde1f88f1730420580061f4068e645ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape615249d-2b", "ovs_interfaceid": "e615249d-2b37-4211-abe1-70015ea3d4cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 816.498744] env[65385]: DEBUG nova.scheduler.client.report [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 816.516906] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527048d2-e5ac-d333-15cc-b4b8cb22618a, 'name': SearchDatastore_Task, 'duration_secs': 0.020875} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.518312] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e603477-98fd-4357-b03a-1ed82a937b2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.525523] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 816.525523] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524d78cb-1dc5-92d9-db53-8e0a167e9b4e" [ 816.525523] env[65385]: _type = "Task" [ 816.525523] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.535689] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524d78cb-1dc5-92d9-db53-8e0a167e9b4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.596105] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "3657ec21-3ec7-44c9-92d1-570655b58f26" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.598575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.598575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "3657ec21-3ec7-44c9-92d1-570655b58f26-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.598575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.598575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.599491] env[65385]: INFO nova.compute.manager [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Terminating instance [ 816.648467] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453717, 'name': PowerOffVM_Task, 'duration_secs': 0.286383} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.648930] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.648930] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 816.649171] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870980', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'name': 'volume-562ed51f-6604-4a40-be52-647787924294', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1ad1e02b-fb82-4a88-9047-be316d590bd2', 'attached_at': '', 'detached_at': '', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'serial': '562ed51f-6604-4a40-be52-647787924294'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 816.649964] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8c5897-eb91-4eb2-a4c6-017403c5f4ae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.669570] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a22e60f-f49e-44f3-9e95-27802993e45e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.678221] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03637bda-e23f-4522-8dac-d8e041745a00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.698948] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa312c0-d20c-41c6-93ed-aebf23034010 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.715119] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] The volume has not been displaced from its original location: [datastore2] volume-562ed51f-6604-4a40-be52-647787924294/volume-562ed51f-6604-4a40-be52-647787924294.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 816.720485] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Reconfiguring VM instance instance-0000002e to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 816.720878] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4992c5f-cb57-4290-8514-0658833fde46 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.741595] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 816.741595] env[65385]: value = "task-4453719" [ 816.741595] env[65385]: _type = "Task" [ 816.741595] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.750313] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453719, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.785350] env[65385]: DEBUG oslo_vmware.api [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453718, 'name': PowerOffVM_Task, 'duration_secs': 0.314117} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.786258] env[65385]: INFO nova.compute.manager [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Took 40.77 seconds to build instance. [ 816.787886] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.788237] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.788886] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2043b40-68e4-4e69-a045-81db0be384e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.792222] env[65385]: DEBUG oslo_concurrency.lockutils [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] Releasing lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 816.793902] env[65385]: DEBUG nova.compute.manager [req-7d80f989-999d-40bb-b90e-a36cf97cb807 req-86b241fa-8954-4a40-a6bc-bfa89c7124dc service nova] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Received event network-vif-deleted-97c1d2c0-12b5-428b-85b4-81b8305d87f9 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 816.878470] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.878732] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.878926] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Deleting the datastore file [datastore1] 483cf486-f0fc-48a4-9db2-970d0f35865a {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.879323] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0913685d-5b43-414b-a7ac-edecdaf08221 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.887603] env[65385]: DEBUG oslo_vmware.api [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 816.887603] env[65385]: value = "task-4453721" [ 816.887603] env[65385]: _type = "Task" [ 816.887603] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.900982] env[65385]: DEBUG oslo_vmware.api [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.967564] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Releasing lock "refresh_cache-af8d01e2-fb86-41c4-99a5-204b30eeda0c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 816.968140] env[65385]: DEBUG nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Instance network_info: |[{"id": "e615249d-2b37-4211-abe1-70015ea3d4cf", "address": "fa:16:3e:a4:42:90", "network": {"id": "109b5b1e-bc16-42c7-a4fe-3f5b85a52cfa", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-704603508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dde1f88f1730420580061f4068e645ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape615249d-2b", "ovs_interfaceid": "e615249d-2b37-4211-abe1-70015ea3d4cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 816.968522] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:42:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7806fe18-2b89-4386-87b1-f22876f82af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e615249d-2b37-4211-abe1-70015ea3d4cf', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.976418] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Creating folder: Project (dde1f88f1730420580061f4068e645ac). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.976887] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-332091cc-8c65-4732-8191-beb23aeec7a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.992284] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Created folder: Project (dde1f88f1730420580061f4068e645ac) in parent group-v870881. [ 816.992529] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Creating folder: Instances. Parent ref: group-v871052. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.992819] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d902d1e-e181-4b56-95cc-7313beb852d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.004015] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.186s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.008404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 30.020s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.008638] env[65385]: DEBUG nova.objects.instance [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 817.012351] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Created folder: Instances in parent group-v871052. [ 817.012619] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 817.012841] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 817.013075] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-673d4fca-3c94-4860-add8-b5aa2fbd7123 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.042094] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524d78cb-1dc5-92d9-db53-8e0a167e9b4e, 'name': SearchDatastore_Task, 'duration_secs': 0.013912} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.043395] env[65385]: INFO nova.scheduler.client.report [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Deleted allocations for instance 649b4e21-fc55-415c-a1f3-ec724397b874 [ 817.051025] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 817.051330] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.051633] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.051633] env[65385]: value = "task-4453724" [ 817.051633] env[65385]: _type = "Task" [ 817.051633] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.052422] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79487d62-68e6-4a66-8bab-b2579c5a1343 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.064397] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453724, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.066719] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 817.066719] env[65385]: value = "task-4453725" [ 817.066719] env[65385]: _type = "Task" [ 817.066719] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.077293] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453725, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.103528] env[65385]: DEBUG nova.compute.manager [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 817.103813] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 817.105089] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857916a9-2fcd-4630-9876-2d0bef5395a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.115522] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 817.115879] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49836d1e-0330-45c2-9709-646477d0cb52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.126308] env[65385]: DEBUG oslo_vmware.api [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 817.126308] env[65385]: value = "task-4453726" [ 817.126308] env[65385]: _type = "Task" [ 817.126308] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.139054] env[65385]: DEBUG oslo_vmware.api [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453726, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.253022] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453719, 'name': ReconfigVM_Task, 'duration_secs': 0.25014} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.253430] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Reconfigured VM instance instance-0000002e to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 817.259362] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-544f8b33-4a0b-4c1c-b083-b0592d663578 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.278161] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 817.278161] env[65385]: value = "task-4453727" [ 817.278161] env[65385]: _type = "Task" [ 817.278161] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.288487] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453727, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.290318] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05f0f228-e1dd-42ae-94f2-6fdf9e6adaae tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "f25cc41e-560e-4cab-8bf0-dac47899d037" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.729s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.345302] env[65385]: DEBUG nova.compute.manager [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Received event network-vif-plugged-e615249d-2b37-4211-abe1-70015ea3d4cf {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 817.345717] env[65385]: DEBUG oslo_concurrency.lockutils [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Acquiring lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 817.346073] env[65385]: DEBUG oslo_concurrency.lockutils [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.346312] env[65385]: DEBUG oslo_concurrency.lockutils [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.346663] env[65385]: DEBUG nova.compute.manager [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] No waiting events found dispatching network-vif-plugged-e615249d-2b37-4211-abe1-70015ea3d4cf {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 817.346916] env[65385]: WARNING nova.compute.manager [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Received unexpected event network-vif-plugged-e615249d-2b37-4211-abe1-70015ea3d4cf for instance with vm_state building and task_state spawning. [ 817.347157] env[65385]: DEBUG nova.compute.manager [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Received event network-changed-e615249d-2b37-4211-abe1-70015ea3d4cf {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 817.347350] env[65385]: DEBUG nova.compute.manager [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Refreshing instance network info cache due to event network-changed-e615249d-2b37-4211-abe1-70015ea3d4cf. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 817.347575] env[65385]: DEBUG oslo_concurrency.lockutils [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Acquiring lock "refresh_cache-af8d01e2-fb86-41c4-99a5-204b30eeda0c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.347751] env[65385]: DEBUG oslo_concurrency.lockutils [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Acquired lock "refresh_cache-af8d01e2-fb86-41c4-99a5-204b30eeda0c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 817.347926] env[65385]: DEBUG nova.network.neutron [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Refreshing network info cache for port e615249d-2b37-4211-abe1-70015ea3d4cf {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 817.409600] env[65385]: DEBUG oslo_vmware.api [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314245} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.411111] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.411111] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.411111] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 817.411111] env[65385]: INFO nova.compute.manager [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Took 1.18 seconds to destroy the instance on the hypervisor. [ 817.411442] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 817.412240] env[65385]: DEBUG nova.compute.manager [-] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 817.412240] env[65385]: DEBUG nova.network.neutron [-] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 817.413080] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.413402] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.413873] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.478811] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.559556] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e1e5ab7c-7776-445f-b66d-ff9b72e0a86d tempest-ListServerFiltersTestJSON-139534771 tempest-ListServerFiltersTestJSON-139534771-project-member] Lock "649b4e21-fc55-415c-a1f3-ec724397b874" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.504s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.571189] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453724, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.587481] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453725, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.647825] env[65385]: DEBUG oslo_vmware.api [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453726, 'name': PowerOffVM_Task, 'duration_secs': 0.319532} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.647998] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 817.648234] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 817.648668] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a897af4-f791-4184-8350-48b6ced4320e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.742934] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 817.743371] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 817.743698] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Deleting the datastore file [datastore1] 3657ec21-3ec7-44c9-92d1-570655b58f26 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 817.744124] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1850a77-866a-4cea-9278-6d910a59f7ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.760349] env[65385]: DEBUG oslo_vmware.api [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for the task: (returnval){ [ 817.760349] env[65385]: value = "task-4453729" [ 817.760349] env[65385]: _type = "Task" [ 817.760349] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.771432] env[65385]: DEBUG oslo_vmware.api [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.794105] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453727, 'name': ReconfigVM_Task, 'duration_secs': 0.209203} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.794864] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-870980', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'name': 'volume-562ed51f-6604-4a40-be52-647787924294', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1ad1e02b-fb82-4a88-9047-be316d590bd2', 'attached_at': '', 'detached_at': '', 'volume_id': '562ed51f-6604-4a40-be52-647787924294', 'serial': '562ed51f-6604-4a40-be52-647787924294'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 817.795223] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 817.796197] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36bb467e-ce27-47de-a7b1-87e27424a09f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.808978] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 817.809305] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a9583ff-1853-46ce-8703-3e9d0d775f15 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.851987] env[65385]: WARNING neutronclient.v2_0.client [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.853033] env[65385]: WARNING openstack [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.853390] env[65385]: WARNING openstack [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.897428] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 817.897720] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 817.897858] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Deleting the datastore file [datastore2] 1ad1e02b-fb82-4a88-9047-be316d590bd2 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 817.898168] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db286c40-f0d3-4fea-8fb8-dbcdeeffe499 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.906834] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for the task: (returnval){ [ 817.906834] env[65385]: value = "task-4453731" [ 817.906834] env[65385]: _type = "Task" [ 817.906834] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.917223] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.978223] env[65385]: WARNING openstack [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.978223] env[65385]: WARNING openstack [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.019850] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0eeb698d-3cdf-4211-9161-44aa49dcc8b0 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 818.025358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.999s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 818.025769] env[65385]: DEBUG nova.objects.instance [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lazy-loading 'resources' on Instance uuid e3248f4a-8985-48b0-84cf-795f00ff21fa {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.063661] env[65385]: WARNING neutronclient.v2_0.client [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.064361] env[65385]: WARNING openstack [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.064709] env[65385]: WARNING openstack [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.081619] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453724, 'name': CreateVM_Task, 'duration_secs': 0.801683} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.082227] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 818.082790] env[65385]: WARNING neutronclient.v2_0.client [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.083205] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.083359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 818.083695] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 818.083981] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-727bb6d1-f832-45b1-a403-f922c71b0263 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.090943] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453725, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693854} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.092631] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.092631] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.092631] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-99bd8148-c7fc-4002-b2db-8e1f9040bc81 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.096799] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 818.096799] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279f4f6-cc97-d588-8b86-4c09cf15dd5f" [ 818.096799] env[65385]: _type = "Task" [ 818.096799] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.102864] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 818.102864] env[65385]: value = "task-4453732" [ 818.102864] env[65385]: _type = "Task" [ 818.102864] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.106612] env[65385]: INFO nova.compute.manager [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Rebuilding instance [ 818.108793] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279f4f6-cc97-d588-8b86-4c09cf15dd5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.119062] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.167338] env[65385]: DEBUG nova.network.neutron [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Updated VIF entry in instance network info cache for port e615249d-2b37-4211-abe1-70015ea3d4cf. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 818.167727] env[65385]: DEBUG nova.network.neutron [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Updating instance_info_cache with network_info: [{"id": "e615249d-2b37-4211-abe1-70015ea3d4cf", "address": "fa:16:3e:a4:42:90", "network": {"id": "109b5b1e-bc16-42c7-a4fe-3f5b85a52cfa", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-704603508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dde1f88f1730420580061f4068e645ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape615249d-2b", "ovs_interfaceid": "e615249d-2b37-4211-abe1-70015ea3d4cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 818.171440] env[65385]: DEBUG nova.compute.manager [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 818.171584] env[65385]: DEBUG nova.compute.manager [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing instance network info cache due to event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 818.172341] env[65385]: DEBUG oslo_concurrency.lockutils [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Acquiring lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.172341] env[65385]: DEBUG oslo_concurrency.lockutils [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Acquired lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 818.172743] env[65385]: DEBUG nova.network.neutron [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 818.178237] env[65385]: DEBUG nova.compute.manager [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 818.179682] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f50e00c-8fc8-44d4-9205-ceebce695b0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.275554] env[65385]: DEBUG oslo_vmware.api [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Task: {'id': task-4453729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188097} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.275867] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 818.276066] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 818.276270] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 818.276446] env[65385]: INFO nova.compute.manager [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Took 1.17 seconds to destroy the instance on the hypervisor. [ 818.276686] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 818.276885] env[65385]: DEBUG nova.compute.manager [-] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 818.277147] env[65385]: DEBUG nova.network.neutron [-] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 818.277345] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.277903] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.278183] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.295913] env[65385]: DEBUG nova.network.neutron [-] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 818.321043] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.418266] env[65385]: DEBUG oslo_vmware.api [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Task: {'id': task-4453731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12372} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.418547] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 818.418643] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 818.418775] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 818.418984] env[65385]: INFO nova.compute.manager [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Took 2.29 seconds to destroy the instance on the hypervisor. [ 818.419258] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 818.419479] env[65385]: DEBUG nova.compute.manager [-] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 818.419587] env[65385]: DEBUG nova.network.neutron [-] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 818.419839] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.420430] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.420712] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.621905] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279f4f6-cc97-d588-8b86-4c09cf15dd5f, 'name': SearchDatastore_Task, 'duration_secs': 0.013613} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.626896] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.627237] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.627509] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.627924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 818.628776] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.628776] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076204} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.631503] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42eb8fd5-7236-4735-a040-aeb7856c927e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.634034] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.635570] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba822c98-b255-4581-be40-b7190a894f78 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.662604] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.664924] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.669069] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bd4c05a-f601-43bb-ac0b-a5f59f08cc26 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.684347] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.684614] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.686868] env[65385]: DEBUG oslo_concurrency.lockutils [req-99f46b4e-6baf-463d-8996-5172da776076 req-c422a2be-7a16-4712-b7d7-1c32544f12f3 service nova] Releasing lock "refresh_cache-af8d01e2-fb86-41c4-99a5-204b30eeda0c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.687493] env[65385]: WARNING neutronclient.v2_0.client [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.688068] env[65385]: WARNING openstack [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.688407] env[65385]: WARNING openstack [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.696409] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b655bd30-af6f-4d00-8221-ea449a842e05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.709575] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 818.709575] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5244482c-5d40-8bfb-f730-8156cdf4a5e7" [ 818.709575] env[65385]: _type = "Task" [ 818.709575] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.709575] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 818.709575] env[65385]: value = "task-4453733" [ 818.709575] env[65385]: _type = "Task" [ 818.709575] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.723690] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5244482c-5d40-8bfb-f730-8156cdf4a5e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.727640] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453733, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.799815] env[65385]: INFO nova.compute.manager [-] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Took 1.39 seconds to deallocate network for instance. [ 819.078239] env[65385]: DEBUG nova.network.neutron [-] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.124418] env[65385]: WARNING openstack [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.124886] env[65385]: WARNING openstack [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.193352] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d4d3e8-0f50-4c2a-a84e-098da4d91915 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.201753] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb215c3-8345-4e89-a9b8-852ffb7fa3f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.208485] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 819.240706] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37d4f136-5e1f-4005-b1a1-f9cbd54ca0fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.253937] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee141bd2-63c4-46a2-ae2a-19dad6a1504c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.284326] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7e8ec2-249d-4318-8530-d1bbe6b01e09 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.289295] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5244482c-5d40-8bfb-f730-8156cdf4a5e7, 'name': SearchDatastore_Task, 'duration_secs': 0.031778} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.289610] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 819.289610] env[65385]: value = "task-4453734" [ 819.289610] env[65385]: _type = "Task" [ 819.289610] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.290393] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453733, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.291750] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b73ef503-ccc3-4a2a-9de5-080972883302 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.306715] env[65385]: DEBUG nova.compute.provider_tree [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.312102] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 819.313492] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 819.313492] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522ed98d-2685-8f84-7119-20e37164cf2e" [ 819.313492] env[65385]: _type = "Task" [ 819.313492] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.317423] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.330755] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522ed98d-2685-8f84-7119-20e37164cf2e, 'name': SearchDatastore_Task, 'duration_secs': 0.012631} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.331252] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 819.331530] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] af8d01e2-fb86-41c4-99a5-204b30eeda0c/af8d01e2-fb86-41c4-99a5-204b30eeda0c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.331856] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f502d767-7181-4fd1-8455-517f86d8439a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.340501] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 819.340501] env[65385]: value = "task-4453735" [ 819.340501] env[65385]: _type = "Task" [ 819.340501] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.352351] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.465610] env[65385]: WARNING neutronclient.v2_0.client [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.465610] env[65385]: WARNING openstack [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.465610] env[65385]: WARNING openstack [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.587950] env[65385]: INFO nova.compute.manager [-] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Took 1.31 seconds to deallocate network for instance. [ 819.593182] env[65385]: DEBUG nova.network.neutron [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updated VIF entry in instance network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 819.593955] env[65385]: DEBUG nova.network.neutron [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updating instance_info_cache with network_info: [{"id": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "address": "fa:16:3e:01:88:e2", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf61ecab0-44", "ovs_interfaceid": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.760893] env[65385]: DEBUG nova.network.neutron [-] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.762371] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453733, 'name': ReconfigVM_Task, 'duration_secs': 0.751757} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.763908] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.764498] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e30fd8e-d62c-43d6-b3d4-6c3a37c48a62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.774317] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 819.774317] env[65385]: value = "task-4453736" [ 819.774317] env[65385]: _type = "Task" [ 819.774317] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.788041] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453736, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.802982] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453734, 'name': PowerOffVM_Task, 'duration_secs': 0.19268} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.803367] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.803986] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 819.805475] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1fc8e14-af0f-4f61-b992-d74a32af5022 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.811151] env[65385]: DEBUG nova.scheduler.client.report [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 819.823750] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 819.825017] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57ee0f53-b118-4567-a239-367c18e7a86f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.853406] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453735, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.857898] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 819.858074] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 819.858255] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleting the datastore file [datastore1] f25cc41e-560e-4cab-8bf0-dac47899d037 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.858583] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-511c58a7-e472-4066-883d-f1161b19bfb0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.870973] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 819.870973] env[65385]: value = "task-4453738" [ 819.870973] env[65385]: _type = "Task" [ 819.870973] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.884205] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.100236] env[65385]: DEBUG oslo_concurrency.lockutils [req-aebcf951-e060-4fea-9c0b-6a3067f0089b req-871fc773-2fad-48bd-aba6-99bc1eaa0760 service nova] Releasing lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.102183] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 820.111343] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521ac185-c5d8-cc40-be0d-75322554c767/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 820.112360] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdc9f00-991d-43d7-80e6-4dfa2b9241be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.120149] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521ac185-c5d8-cc40-be0d-75322554c767/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 820.120388] env[65385]: ERROR oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521ac185-c5d8-cc40-be0d-75322554c767/disk-0.vmdk due to incomplete transfer. [ 820.120638] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-667e51fd-306c-4a4d-8cbf-9bf977990672 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.129523] env[65385]: DEBUG oslo_vmware.rw_handles [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521ac185-c5d8-cc40-be0d-75322554c767/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 820.129745] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Uploaded image 5e8408d4-fb50-4b8f-95d6-5374d829517e to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 820.133374] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 820.133374] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-580ba87f-2ad8-449a-a4d8-fc683d1770e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.143338] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 820.143338] env[65385]: value = "task-4453739" [ 820.143338] env[65385]: _type = "Task" [ 820.143338] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.150869] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453739, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.265099] env[65385]: INFO nova.compute.manager [-] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Took 1.85 seconds to deallocate network for instance. [ 820.289083] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453736, 'name': Rename_Task, 'duration_secs': 0.291947} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.289460] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.291046] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a65ea568-da91-41aa-9030-49cb522b279c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.300897] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 820.300897] env[65385]: value = "task-4453740" [ 820.300897] env[65385]: _type = "Task" [ 820.300897] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.314930] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.319341] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.298s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.323687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.093s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.328355] env[65385]: INFO nova.compute.claims [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.356607] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453735, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523417} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.357226] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] af8d01e2-fb86-41c4-99a5-204b30eeda0c/af8d01e2-fb86-41c4-99a5-204b30eeda0c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.357805] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.358058] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0c9c3c3-f3df-4c53-bd4d-03688a75ef8a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.361206] env[65385]: INFO nova.scheduler.client.report [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleted allocations for instance e3248f4a-8985-48b0-84cf-795f00ff21fa [ 820.371608] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 820.371608] env[65385]: value = "task-4453741" [ 820.371608] env[65385]: _type = "Task" [ 820.371608] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.391914] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113386} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.391914] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453741, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.392222] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.392416] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.392590] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.623811] env[65385]: DEBUG nova.compute.manager [req-12c6b37e-5ab8-40f1-993f-5ecbe44f8701 req-4a1954bc-a10d-4fc7-acdd-be73ecd6ad5e service nova] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Received event network-vif-deleted-828ba549-d28e-4ae9-bbf6-7050becb6e96 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 820.657134] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453739, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.813433] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453740, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.837687] env[65385]: INFO nova.compute.manager [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Took 0.57 seconds to detach 1 volumes for instance. [ 820.840352] env[65385]: DEBUG nova.compute.manager [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Deleting volume: 562ed51f-6604-4a40-be52-647787924294 {{(pid=65385) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3320}} [ 820.868811] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f58485e0-2981-4d5a-8918-4c9b60947b72 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "e3248f4a-8985-48b0-84cf-795f00ff21fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.569s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.887032] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453741, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090892} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.889164] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.889164] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480ce415-5d68-4045-afed-1857a28c42b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.920249] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] af8d01e2-fb86-41c4-99a5-204b30eeda0c/af8d01e2-fb86-41c4-99a5-204b30eeda0c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.921751] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcfe1fae-5c89-472e-9380-90325ecfe9de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.944132] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 820.944132] env[65385]: value = "task-4453742" [ 820.944132] env[65385]: _type = "Task" [ 820.944132] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.956264] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.158057] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453739, 'name': Destroy_Task, 'duration_secs': 0.887728} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.160553] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Destroyed the VM [ 821.160553] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 821.162054] env[65385]: DEBUG nova.compute.manager [req-2217d27c-d26d-4c21-be7c-32846bc46b93 req-54cd67df-fc3a-495b-ae57-7182d0d77972 service nova] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Received event network-vif-deleted-4090f583-8481-4e0c-9a39-3a393fb0ea72 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 821.162473] env[65385]: DEBUG nova.compute.manager [req-2217d27c-d26d-4c21-be7c-32846bc46b93 req-54cd67df-fc3a-495b-ae57-7182d0d77972 service nova] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Received event network-vif-deleted-94ad8d4c-044d-465c-a0f7-b58c7f70cc14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 821.162818] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-39816624-de96-4f18-a1e4-99855cbc5d0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.172146] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 821.172146] env[65385]: value = "task-4453744" [ 821.172146] env[65385]: _type = "Task" [ 821.172146] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.189820] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453744, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.315252] env[65385]: DEBUG oslo_vmware.api [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453740, 'name': PowerOnVM_Task, 'duration_secs': 0.896922} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.315610] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.315843] env[65385]: INFO nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Took 9.63 seconds to spawn the instance on the hypervisor. [ 821.316024] env[65385]: DEBUG nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 821.316855] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897aa797-52e3-48b9-8a7c-48001ade5259 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.403629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.456456] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453742, 'name': ReconfigVM_Task, 'duration_secs': 0.344397} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.458894] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Reconfigured VM instance instance-0000003c to attach disk [datastore2] af8d01e2-fb86-41c4-99a5-204b30eeda0c/af8d01e2-fb86-41c4-99a5-204b30eeda0c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.460080] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e9a2ff9-a585-427e-819d-b7dd4e5f2201 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.470943] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 821.471336] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 821.471501] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 821.471736] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 821.471948] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 821.472158] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 821.472443] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.473149] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 821.473149] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 821.473149] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 821.473343] env[65385]: DEBUG nova.virt.hardware [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 821.475842] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40875fdf-2ce2-4cba-b906-8f8ff75faf8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.478789] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 821.478789] env[65385]: value = "task-4453745" [ 821.478789] env[65385]: _type = "Task" [ 821.478789] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.492929] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-465e5fe6-263e-44ee-b335-4629ba87c935 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.502267] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453745, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.513384] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.519701] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 821.523352] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.524282] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f312e8c7-3de8-4c1b-b940-5164099a8c0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.543467] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.543467] env[65385]: value = "task-4453746" [ 821.543467] env[65385]: _type = "Task" [ 821.543467] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.552967] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453746, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.683417] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453744, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.839532] env[65385]: INFO nova.compute.manager [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Took 43.25 seconds to build instance. [ 821.960493] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9843e82a-848f-4f85-8282-3619e9bc1043 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.969408] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf362e8-83b7-4037-bea8-9482d2c86c60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.006393] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a63168e-ec42-4449-8da5-773e1363156c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.021290] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453745, 'name': Rename_Task, 'duration_secs': 0.170207} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.022976] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 822.023291] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff14008f-da54-4660-9bd6-4cc0fb36fe56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.026499] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6260b4ee-6ca4-403c-b874-66386e81341e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.044547] env[65385]: DEBUG nova.compute.provider_tree [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.054789] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 822.054789] env[65385]: value = "task-4453747" [ 822.054789] env[65385]: _type = "Task" [ 822.054789] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.063604] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453746, 'name': CreateVM_Task, 'duration_secs': 0.512948} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.064197] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.064663] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.064810] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.065150] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 822.065839] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd029d3b-ed9a-4127-8004-fc52e4719035 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.071483] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453747, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.075089] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 822.075089] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526b696d-7829-94bd-d702-aead8d94c434" [ 822.075089] env[65385]: _type = "Task" [ 822.075089] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.086180] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526b696d-7829-94bd-d702-aead8d94c434, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.186985] env[65385]: DEBUG oslo_vmware.api [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453744, 'name': RemoveSnapshot_Task, 'duration_secs': 0.600226} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.187331] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 822.187555] env[65385]: INFO nova.compute.manager [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Took 16.86 seconds to snapshot the instance on the hypervisor. [ 822.340334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae4bc535-54c7-438b-894d-2418d088d001 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.119s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.556883] env[65385]: DEBUG nova.scheduler.client.report [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 822.574939] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453747, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.589339] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526b696d-7829-94bd-d702-aead8d94c434, 'name': SearchDatastore_Task, 'duration_secs': 0.036785} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.589637] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.589864] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.590106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.590271] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.590445] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.590720] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53f4e1d4-ed27-45df-8598-68d4be1db4b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.605556] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.605556] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.605706] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-947c07aa-dd29-454a-a8c7-4783e941b109 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.614830] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 822.614830] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cd843a-862c-de7a-db75-cd0b6b792f3e" [ 822.614830] env[65385]: _type = "Task" [ 822.614830] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.626798] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cd843a-862c-de7a-db75-cd0b6b792f3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.691888] env[65385]: DEBUG nova.compute.manager [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Instance disappeared during snapshot {{(pid=65385) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4635}} [ 822.708114] env[65385]: DEBUG nova.compute.manager [None req-dc8645ed-b3dc-4c78-8da3-c80ec7b93872 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image not found during clean up 5e8408d4-fb50-4b8f-95d6-5374d829517e {{(pid=65385) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4641}} [ 822.729500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "c5d00bb5-d869-466f-826e-61622e57959f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.729758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "c5d00bb5-d869-466f-826e-61622e57959f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.730869] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "c5d00bb5-d869-466f-826e-61622e57959f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.730869] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "c5d00bb5-d869-466f-826e-61622e57959f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.730869] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "c5d00bb5-d869-466f-826e-61622e57959f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.737116] env[65385]: INFO nova.compute.manager [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Terminating instance [ 823.069488] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.746s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.069488] env[65385]: DEBUG nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 823.078876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.024s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.078982] env[65385]: DEBUG nova.objects.instance [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lazy-loading 'resources' on Instance uuid 2f66a03b-874b-4bee-9694-49813ecb0c8a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 823.080623] env[65385]: DEBUG oslo_vmware.api [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453747, 'name': PowerOnVM_Task, 'duration_secs': 0.640924} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.088024] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 823.088024] env[65385]: INFO nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Took 8.61 seconds to spawn the instance on the hypervisor. [ 823.088024] env[65385]: DEBUG nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 823.088024] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15184828-1fbe-4fa7-abb4-f9f07acf41a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.125750] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cd843a-862c-de7a-db75-cd0b6b792f3e, 'name': SearchDatastore_Task, 'duration_secs': 0.013848} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.126663] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7bdfd24-b185-4ee2-812b-a5a4c8b3fc10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.134515] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 823.134515] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279a2d6-75ad-c1d6-814d-e743c2747549" [ 823.134515] env[65385]: _type = "Task" [ 823.134515] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.146256] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279a2d6-75ad-c1d6-814d-e743c2747549, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.245024] env[65385]: DEBUG nova.compute.manager [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 823.245297] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.246294] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01c7450-d034-4836-9ca4-9c1d01944d48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.256018] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 823.256224] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db485082-2c8f-4239-bb0d-4fe7bd8918d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.353503] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 823.353662] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 823.353843] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleting the datastore file [datastore2] c5d00bb5-d869-466f-826e-61622e57959f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.355872] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fc8e356-ab79-47bb-862d-5753cd20565a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.366458] env[65385]: DEBUG oslo_vmware.api [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 823.366458] env[65385]: value = "task-4453749" [ 823.366458] env[65385]: _type = "Task" [ 823.366458] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.386380] env[65385]: DEBUG oslo_vmware.api [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.587632] env[65385]: DEBUG nova.compute.utils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 823.593146] env[65385]: DEBUG nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 823.593146] env[65385]: DEBUG nova.network.neutron [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 823.593297] env[65385]: WARNING neutronclient.v2_0.client [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.593612] env[65385]: WARNING neutronclient.v2_0.client [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.594213] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.594550] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.617348] env[65385]: INFO nova.compute.manager [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Took 40.91 seconds to build instance. [ 823.649369] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279a2d6-75ad-c1d6-814d-e743c2747549, 'name': SearchDatastore_Task, 'duration_secs': 0.014719} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.649674] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 823.650065] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.650257] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fdbfd2d-e720-401a-9708-6310f94e5cf0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.659216] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 823.659216] env[65385]: value = "task-4453750" [ 823.659216] env[65385]: _type = "Task" [ 823.659216] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.672692] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453750, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.757302] env[65385]: DEBUG nova.policy [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d5103a5769a4296973351194f49fed3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '754b1facaaa14501b2204c98e1d7a5a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 823.886244] env[65385]: DEBUG oslo_vmware.api [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.240171} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.886749] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.886749] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.886863] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.887041] env[65385]: INFO nova.compute.manager [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Took 0.64 seconds to destroy the instance on the hypervisor. [ 823.887287] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 823.888317] env[65385]: DEBUG nova.compute.manager [-] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 823.888317] env[65385]: DEBUG nova.network.neutron [-] [instance: c5d00bb5-d869-466f-826e-61622e57959f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 823.888317] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.888475] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.888657] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.105519] env[65385]: DEBUG nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 824.122411] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ad1a8bf-439d-459d-a71c-f7332ddcf6fe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.541s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 824.158222] env[65385]: DEBUG nova.compute.manager [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 824.158222] env[65385]: DEBUG nova.compute.manager [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing instance network info cache due to event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 824.158505] env[65385]: DEBUG oslo_concurrency.lockutils [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Acquiring lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.158548] env[65385]: DEBUG oslo_concurrency.lockutils [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Acquired lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.158705] env[65385]: DEBUG nova.network.neutron [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 824.175257] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453750, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.180185] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.181267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.188685] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.323122] env[65385]: DEBUG nova.network.neutron [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Successfully created port: d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 824.533553] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c31a69-dea9-43da-9e0f-c3ffcc30d72e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.545724] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6df6f67-0712-4f26-866f-38b338919e76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.583405] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c084899d-7b43-45f9-97d9-544a856ef848 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.592797] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc57744-7271-4d79-b585-a43b0aee8c32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.610976] env[65385]: DEBUG nova.compute.provider_tree [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.636776] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.636776] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.661644] env[65385]: WARNING neutronclient.v2_0.client [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.662810] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.662810] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.682203] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453750, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.682506] env[65385]: DEBUG nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 825.112776] env[65385]: DEBUG nova.scheduler.client.report [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 825.124153] env[65385]: DEBUG nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 825.140738] env[65385]: DEBUG nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 825.156028] env[65385]: DEBUG nova.network.neutron [-] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 825.170214] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 825.170419] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 825.170778] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 825.170778] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 825.170922] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 825.172044] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 825.172044] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.172044] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 825.172044] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 825.172044] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 825.172044] env[65385]: DEBUG nova.virt.hardware [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 825.173591] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0612e673-bf41-4212-933f-5b5624c36f19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.197370] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453750, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.202054] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ecfc68-2c3a-4a92-97ea-4f7db3a855f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.236567] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 825.540888] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.541293] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.621028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.541s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.622595] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.863s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 825.622893] env[65385]: DEBUG nova.objects.instance [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lazy-loading 'resources' on Instance uuid 34730159-3bec-4ad5-b85e-0f67998d6001 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.661239] env[65385]: INFO nova.compute.manager [-] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Took 1.77 seconds to deallocate network for instance. [ 825.668717] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 825.690223] env[65385]: INFO nova.scheduler.client.report [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Deleted allocations for instance 2f66a03b-874b-4bee-9694-49813ecb0c8a [ 825.701175] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453750, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.680068} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.705330] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.705708] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.706160] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5918791-77e1-404e-aa47-adcbf63f584e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.723026] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 825.723026] env[65385]: value = "task-4453751" [ 825.723026] env[65385]: _type = "Task" [ 825.723026] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.736084] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453751, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.792616] env[65385]: WARNING neutronclient.v2_0.client [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.793653] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.794236] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.824133] env[65385]: DEBUG nova.compute.manager [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Received event network-changed-fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 825.824378] env[65385]: DEBUG nova.compute.manager [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Refreshing instance network info cache due to event network-changed-fe815bb4-491a-42f5-a5d3-2eba8997444a. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 825.824604] env[65385]: DEBUG oslo_concurrency.lockutils [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Acquiring lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.824718] env[65385]: DEBUG oslo_concurrency.lockutils [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Acquired lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.824868] env[65385]: DEBUG nova.network.neutron [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Refreshing network info cache for port fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 825.973346] env[65385]: DEBUG nova.network.neutron [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updated VIF entry in instance network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 825.973686] env[65385]: DEBUG nova.network.neutron [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updating instance_info_cache with network_info: [{"id": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "address": "fa:16:3e:01:88:e2", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf61ecab0-44", "ovs_interfaceid": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 826.179823] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.216318] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cdb68e8f-84a9-4ed9-8249-cd9498d41f47 tempest-AttachInterfacesUnderV243Test-143948563 tempest-AttachInterfacesUnderV243Test-143948563-project-member] Lock "2f66a03b-874b-4bee-9694-49813ecb0c8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.701s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.239143] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453751, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128418} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.240587] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.242472] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74251eb-865a-47b7-b2a4-77a02781d1e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.288222] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.293804] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c3a481b-cb80-4941-91a9-1cfce7c4980c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.315775] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "ad29afd9-3e52-4fac-977f-91774acad7d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.315775] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.316112] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.316381] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.317316] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.322258] env[65385]: INFO nova.compute.manager [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Terminating instance [ 826.328632] env[65385]: WARNING neutronclient.v2_0.client [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 826.329624] env[65385]: WARNING openstack [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.329961] env[65385]: WARNING openstack [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.339414] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 826.339414] env[65385]: value = "task-4453752" [ 826.339414] env[65385]: _type = "Task" [ 826.339414] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.340403] env[65385]: DEBUG nova.network.neutron [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Successfully updated port: d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 826.355060] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.477032] env[65385]: DEBUG oslo_concurrency.lockutils [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Releasing lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 826.477032] env[65385]: DEBUG nova.compute.manager [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Received event network-changed-fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 826.477032] env[65385]: DEBUG nova.compute.manager [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Refreshing instance network info cache due to event network-changed-fe815bb4-491a-42f5-a5d3-2eba8997444a. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 826.477329] env[65385]: DEBUG oslo_concurrency.lockutils [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Acquiring lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.479145] env[65385]: WARNING openstack [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.479515] env[65385]: WARNING openstack [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.576412] env[65385]: WARNING neutronclient.v2_0.client [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 826.576961] env[65385]: WARNING openstack [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.577356] env[65385]: WARNING openstack [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.674535] env[65385]: DEBUG nova.network.neutron [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updated VIF entry in instance network info cache for port fe815bb4-491a-42f5-a5d3-2eba8997444a. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 826.674884] env[65385]: DEBUG nova.network.neutron [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updating instance_info_cache with network_info: [{"id": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "address": "fa:16:3e:68:9f:df", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe815bb4-49", "ovs_interfaceid": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 826.777264] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d13ef3-1d4a-48db-b755-87d7f6c7417e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.785723] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4c3bfe-edcb-4c9e-afbb-775f3128dd99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.822303] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e16851-621e-4380-8ab2-78fa1ef6c993 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.827299] env[65385]: DEBUG nova.compute.manager [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 826.827574] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 826.828674] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1619c49a-920c-4bde-b93b-20f0209dcf25 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.839937] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03417178-44dd-4150-b1c9-e7d5cc2dc260 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.844606] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 826.845034] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d6afc7d-69aa-4225-b1c6-d4d53cf76182 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.850040] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.850262] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 826.850486] env[65385]: DEBUG nova.network.neutron [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 826.862105] env[65385]: DEBUG nova.compute.provider_tree [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.870293] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.870631] env[65385]: DEBUG oslo_vmware.api [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 826.870631] env[65385]: value = "task-4453753" [ 826.870631] env[65385]: _type = "Task" [ 826.870631] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.882451] env[65385]: DEBUG oslo_vmware.api [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.032535] env[65385]: DEBUG nova.compute.manager [req-498514af-d07c-4168-9716-2b88f035a3dd req-74034bd3-69fa-43f9-b886-a7fe4929db4d service nova] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Received event network-vif-deleted-3617c8a9-2863-4483-86be-36b932f830a3 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 827.178211] env[65385]: DEBUG oslo_concurrency.lockutils [req-6be885f7-8d37-4472-89cf-2ad6e60dca5b req-da5b0c55-d1ee-4458-82af-41a3a0cd40ed service nova] Releasing lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 827.179111] env[65385]: DEBUG oslo_concurrency.lockutils [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Acquired lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 827.179111] env[65385]: DEBUG nova.network.neutron [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Refreshing network info cache for port fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 827.356296] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453752, 'name': ReconfigVM_Task, 'duration_secs': 0.727269} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.356560] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Reconfigured VM instance instance-00000039 to attach disk [datastore2] f25cc41e-560e-4cab-8bf0-dac47899d037/f25cc41e-560e-4cab-8bf0-dac47899d037.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.357270] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5f194b8-37b4-4016-9cfd-6ad90c3c80d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.367642] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.367642] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.378331] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 827.378331] env[65385]: value = "task-4453754" [ 827.378331] env[65385]: _type = "Task" [ 827.378331] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.381031] env[65385]: DEBUG nova.scheduler.client.report [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 827.401020] env[65385]: DEBUG oslo_vmware.api [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453753, 'name': PowerOffVM_Task, 'duration_secs': 0.238854} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.401340] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453754, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.401571] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.401732] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 827.401988] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47b1bc29-be92-4773-8fed-32ea826acc1c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.438023] env[65385]: DEBUG nova.network.neutron [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 827.462682] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.463450] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.477502] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 827.477704] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 827.478225] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Deleting the datastore file [datastore1] ad29afd9-3e52-4fac-977f-91774acad7d0 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.478225] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52458c42-3285-4621-8c32-287beea1e97c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.486975] env[65385]: DEBUG oslo_vmware.api [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 827.486975] env[65385]: value = "task-4453756" [ 827.486975] env[65385]: _type = "Task" [ 827.486975] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.498539] env[65385]: DEBUG oslo_vmware.api [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.553717] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "9908d395-545d-4caf-9757-320d6253d61e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.553908] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "9908d395-545d-4caf-9757-320d6253d61e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.631514] env[65385]: WARNING neutronclient.v2_0.client [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 827.632762] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.632762] env[65385]: WARNING openstack [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.684902] env[65385]: WARNING neutronclient.v2_0.client [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 827.684902] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.685036] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.886335] env[65385]: DEBUG nova.network.neutron [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Updating instance_info_cache with network_info: [{"id": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "address": "fa:16:3e:f3:7a:17", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15afa85-05", "ovs_interfaceid": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 827.895157] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.273s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.898549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 29.259s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.908277] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453754, 'name': Rename_Task, 'duration_secs': 0.17466} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.909693] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 827.910219] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94b35343-cd30-4e9e-9a4b-9ce2303b326f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.922173] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 827.922173] env[65385]: value = "task-4453757" [ 827.922173] env[65385]: _type = "Task" [ 827.922173] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.927774] env[65385]: DEBUG nova.compute.manager [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 827.928907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fb2c1e-325d-4f28-8a5c-1c3b13f1679e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.939923] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453757, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.975346] env[65385]: INFO nova.scheduler.client.report [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted allocations for instance 34730159-3bec-4ad5-b85e-0f67998d6001 [ 827.985977] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.985977] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.002056] env[65385]: DEBUG oslo_vmware.api [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453756, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.46946} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.003065] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.003065] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 828.003065] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.004130] env[65385]: INFO nova.compute.manager [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Took 1.18 seconds to destroy the instance on the hypervisor. [ 828.004130] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 828.004130] env[65385]: DEBUG nova.compute.manager [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 828.004130] env[65385]: DEBUG nova.network.neutron [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 828.004664] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.004764] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.005043] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.058063] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.059469] env[65385]: DEBUG nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 828.088528] env[65385]: WARNING neutronclient.v2_0.client [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.088528] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.089239] env[65385]: WARNING openstack [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.330529] env[65385]: DEBUG nova.network.neutron [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updated VIF entry in instance network info cache for port fe815bb4-491a-42f5-a5d3-2eba8997444a. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 828.330945] env[65385]: DEBUG nova.network.neutron [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updating instance_info_cache with network_info: [{"id": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "address": "fa:16:3e:68:9f:df", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe815bb4-49", "ovs_interfaceid": "fe815bb4-491a-42f5-a5d3-2eba8997444a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 828.389187] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.390311] env[65385]: DEBUG nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Instance network_info: |[{"id": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "address": "fa:16:3e:f3:7a:17", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15afa85-05", "ovs_interfaceid": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 828.390823] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:7a:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd15afa85-053b-49b7-a9df-eb592c2ba28d', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.401325] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 828.401325] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.401325] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-010200bd-50a0-4027-988b-fa08842b4a05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.438997] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.438997] env[65385]: value = "task-4453758" [ 828.438997] env[65385]: _type = "Task" [ 828.438997] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.445491] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453757, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.457094] env[65385]: INFO nova.compute.manager [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] instance snapshotting [ 828.460200] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453758, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.462053] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc7e27e-a959-4627-a837-47b8fdd96026 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.492121] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544921d9-7696-46db-8c04-13d632457f56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.496697] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7d413ff0-1478-4d49-82d8-59d2b0a89d4c tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "34730159-3bec-4ad5-b85e-0f67998d6001" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.782s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.590218] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.834408] env[65385]: DEBUG oslo_concurrency.lockutils [req-93f3061d-e56d-4e61-93bd-450c9369dbd2 req-61cd8003-c064-45f7-b529-8cfa066dfd93 service nova] Releasing lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.834771] env[65385]: DEBUG nova.network.neutron [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 828.934191] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Applying migration context for instance 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22 as it has an incoming, in-progress migration 40af7784-05e8-4465-ba0d-2e13af06eb72. Migration status is reverting {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 828.936718] env[65385]: INFO nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating resource usage from migration 40af7784-05e8-4465-ba0d-2e13af06eb72 [ 828.937167] env[65385]: INFO nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating resource usage from migration 73856ca4-6bf7-447d-9523-fd6a10d8cb2a [ 828.952615] env[65385]: DEBUG oslo_vmware.api [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453757, 'name': PowerOnVM_Task, 'duration_secs': 0.783797} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.953417] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.953616] env[65385]: DEBUG nova.compute.manager [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 828.957460] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1088b4f-44cb-4432-9ec8-edbb6f42290c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.961208] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453758, 'name': CreateVM_Task, 'duration_secs': 0.484974} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.961482] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.962027] env[65385]: WARNING neutronclient.v2_0.client [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.962420] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.962541] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.962843] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 828.963194] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e642f8a4-03de-47b4-9ecf-349e356068d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.971286] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0115dde8-b0aa-4960-a452-ced6587bc567 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.971464] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 59450963-d83f-46e8-8c13-05d4f1818c64 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.971586] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance eb9d0510-b453-4695-9e1d-731217b9f8ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.971700] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 42b9edf6-1873-49c3-8074-8eef654ac371 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.971817] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 75ba706f-2e72-4f84-b02f-db4381951e77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.971997] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 59f26752-75ab-45e5-b0a9-bbcef44ecd7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.972159] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0673a0d2-76ac-4cd5-8cc3-b8596877c641 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.972285] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 3657ec21-3ec7-44c9-92d1-570655b58f26 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.972396] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 90c36f67-0c25-4487-a5e4-89827290953b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.972508] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 483cf486-f0fc-48a4-9db2-970d0f35865a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.972616] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 7fcbdc79-688c-479f-94e2-f4542abe714b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 1ad1e02b-fb82-4a88-9047-be316d590bd2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b702ba27-bbc7-4030-a96d-30e1ed656a9b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance f521e37d-aa86-4533-9fbe-7d040ed90e53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance c5d00bb5-d869-466f-826e-61622e57959f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.974549] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 98430d7e-626e-4532-802e-38f7fc583ae3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.974549] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance e9fd0c9f-5795-44df-9dcb-982b7722ae69 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.974549] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance ab8df1d8-615d-4083-bdf4-25f679a64982 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 65f9bc55-7bab-43b0-a974-eb5080389b7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 26a2d464-0aab-463f-8265-948a4f0ff188 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance ad29afd9-3e52-4fac-977f-91774acad7d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance f25cc41e-560e-4cab-8bf0-dac47899d037 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0c345794-036c-45d1-985e-cf5c116bd873 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.974549] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Migration 40af7784-05e8-4465-ba0d-2e13af06eb72 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 828.975543] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.975543] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance af8d01e2-fb86-41c4-99a5-204b30eeda0c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.975543] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b1cca536-e1d9-43f8-8a5c-be3981c106e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 828.979187] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 828.979187] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5258321f-87db-e039-a372-9a771f457f1c" [ 828.979187] env[65385]: _type = "Task" [ 828.979187] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.990604] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5258321f-87db-e039-a372-9a771f457f1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.006257] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 829.006585] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-83258eda-a820-4535-a17e-127769ff3ae7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.017019] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 829.017019] env[65385]: value = "task-4453759" [ 829.017019] env[65385]: _type = "Task" [ 829.017019] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.033212] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453759, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.338418] env[65385]: INFO nova.compute.manager [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Took 1.33 seconds to deallocate network for instance. [ 829.407364] env[65385]: DEBUG nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Received event network-vif-plugged-d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 829.407608] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Acquiring lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.407801] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.407919] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.408085] env[65385]: DEBUG nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] No waiting events found dispatching network-vif-plugged-d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 829.408275] env[65385]: WARNING nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Received unexpected event network-vif-plugged-d15afa85-053b-49b7-a9df-eb592c2ba28d for instance with vm_state building and task_state spawning. [ 829.408442] env[65385]: DEBUG nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Received event network-changed-d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 829.408713] env[65385]: DEBUG nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Refreshing instance network info cache due to event network-changed-d15afa85-053b-49b7-a9df-eb592c2ba28d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 829.408945] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Acquiring lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.409120] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Acquired lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.409322] env[65385]: DEBUG nova.network.neutron [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Refreshing network info cache for port d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 829.477418] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.480538] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance c18b5bde-1a17-4954-a2c5-582815ebfec8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 829.480538] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Migration 73856ca4-6bf7-447d-9523-fd6a10d8cb2a is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 829.480538] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 89f3c815-8671-47ce-9e74-bf6e652bb3c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 829.491859] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5258321f-87db-e039-a372-9a771f457f1c, 'name': SearchDatastore_Task, 'duration_secs': 0.019104} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.492923] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.492923] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.492923] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.492923] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.492923] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.493229] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-104e5401-7353-446e-a22d-0e975084e4ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.505435] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.505620] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.506376] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-386471b6-8d0a-4119-98e0-5504c3e836f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.512967] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 829.512967] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b28f36-7c33-3008-473c-d138e8173310" [ 829.512967] env[65385]: _type = "Task" [ 829.512967] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.524819] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b28f36-7c33-3008-473c-d138e8173310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.530845] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453759, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.772104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.772367] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.772569] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.772747] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.772909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.775045] env[65385]: INFO nova.compute.manager [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Terminating instance [ 829.845921] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.914464] env[65385]: WARNING neutronclient.v2_0.client [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.915358] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.915648] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.987373] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 830.028949] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453759, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.032448] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b28f36-7c33-3008-473c-d138e8173310, 'name': SearchDatastore_Task, 'duration_secs': 0.016983} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.033365] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-253bb124-d552-4b04-b87e-9ecb65346e3f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.040205] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 830.040205] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261fc27-a97d-4cb1-71ae-4e73ca8c4007" [ 830.040205] env[65385]: _type = "Task" [ 830.040205] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.050789] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261fc27-a97d-4cb1-71ae-4e73ca8c4007, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.144690] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.148133] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.268675] env[65385]: WARNING neutronclient.v2_0.client [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.268675] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.268675] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.280821] env[65385]: DEBUG nova.compute.manager [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 830.281586] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 830.282404] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee5af86-20ee-400e-9f0e-7a6fbd50bd53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.293637] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.294168] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e8ce2e3-9d2b-433b-825f-b4039c0c919b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.304281] env[65385]: DEBUG oslo_vmware.api [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 830.304281] env[65385]: value = "task-4453760" [ 830.304281] env[65385]: _type = "Task" [ 830.304281] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.314521] env[65385]: DEBUG oslo_vmware.api [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.492933] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0c6206ee-c787-4cd1-9289-1b2620dc4c5f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 830.499253] env[65385]: DEBUG nova.network.neutron [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Updated VIF entry in instance network info cache for port d15afa85-053b-49b7-a9df-eb592c2ba28d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 830.499253] env[65385]: DEBUG nova.network.neutron [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Updating instance_info_cache with network_info: [{"id": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "address": "fa:16:3e:f3:7a:17", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15afa85-05", "ovs_interfaceid": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.534878] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453759, 'name': CreateSnapshot_Task, 'duration_secs': 1.15179} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.534959] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 830.535754] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6637a4cf-8e82-4630-803a-ea46844f8376 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.558899] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261fc27-a97d-4cb1-71ae-4e73ca8c4007, 'name': SearchDatastore_Task, 'duration_secs': 0.019688} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.559222] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.559483] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b1cca536-e1d9-43f8-8a5c-be3981c106e7/b1cca536-e1d9-43f8-8a5c-be3981c106e7.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.560268] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d269b193-0497-4443-a7d7-0f72abc1c9ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.569546] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 830.569546] env[65385]: value = "task-4453761" [ 830.569546] env[65385]: _type = "Task" [ 830.569546] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.584208] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.815291] env[65385]: DEBUG oslo_vmware.api [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453760, 'name': PowerOffVM_Task, 'duration_secs': 0.259144} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.815621] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.815789] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 830.816147] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c1b2bf4-49ff-472b-85c9-95486a62daec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.830805] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "370b6aa8-2337-48e1-8102-611ce4494a55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.831054] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "370b6aa8-2337-48e1-8102-611ce4494a55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.898861] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.898861] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.899202] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleting the datastore file [datastore2] 0673a0d2-76ac-4cd5-8cc3-b8596877c641 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.899367] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-618b2dad-cc56-4849-91b7-851a59e6a087 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.912784] env[65385]: DEBUG oslo_vmware.api [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 830.912784] env[65385]: value = "task-4453763" [ 830.912784] env[65385]: _type = "Task" [ 830.912784] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.926468] env[65385]: DEBUG oslo_vmware.api [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453763, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.000411] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 9908d395-545d-4caf-9757-320d6253d61e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 831.000775] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 20 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 831.000984] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4480MB phys_disk=100GB used_disk=21GB total_vcpus=48 used_vcpus=20 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '19', 'num_vm_active': '16', 'num_task_None': '13', 'num_os_type_None': '19', 'num_proj_99173b1f52aa42f9b0db92366db8c411': '2', 'io_workload': '3', 'num_proj_a9a1bf7c2dad4808a3782a3fba8a25fc': '3', 'num_vm_stopped': '1', 'num_task_resize_prep': '1', 'num_proj_1598307c304d46a884a88b3eb5f5d7dc': '1', 'num_proj_60de0d1c162342209795a180391661f6': '1', 'num_proj_cb1ecd79d8e44f6096ca69a5238851b0': '1', 'num_proj_5a3e5ed6b95f4282b1ef29d17fa58975': '2', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_2de8098800694299aae5aa2f59f405bb': '1', 'num_proj_9e412a22bd7b412996ee52adbea024bf': '2', 'num_proj_04be7e71aa654488a30f2a7afccc9f31': '2', 'num_proj_2ba61d77cc2d4fe1ba3c03466d5985d9': '1', 'num_task_deleting': '1', 'num_task_rebuild_spawning': '1', 'num_proj_647f4e3157c34a8f9a64cb8ba8b72da6': '1', 'num_task_image_snapshot': '1', 'num_proj_dde1f88f1730420580061f4068e645ac': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_754b1facaaa14501b2204c98e1d7a5a8': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 831.004385] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Releasing lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 831.004636] env[65385]: DEBUG nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 831.004859] env[65385]: DEBUG nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing instance network info cache due to event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 831.005266] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Acquiring lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.005449] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Acquired lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.005625] env[65385]: DEBUG nova.network.neutron [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 831.069140] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 831.069822] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5d025467-0ef3-4ffa-b0eb-965fb81b893a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.085754] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 831.085754] env[65385]: value = "task-4453764" [ 831.085754] env[65385]: _type = "Task" [ 831.085754] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.090156] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453761, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.104933] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453764, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.190075] env[65385]: DEBUG nova.compute.manager [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 831.190228] env[65385]: DEBUG nova.compute.manager [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing instance network info cache due to event network-changed-f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 831.190417] env[65385]: DEBUG oslo_concurrency.lockutils [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Acquiring lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.333712] env[65385]: DEBUG nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 831.430112] env[65385]: DEBUG oslo_vmware.api [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453763, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.460782} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.438259] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.438475] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.438661] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.438844] env[65385]: INFO nova.compute.manager [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Took 1.16 seconds to destroy the instance on the hypervisor. [ 831.439123] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 831.440429] env[65385]: DEBUG nova.compute.manager [-] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 831.440429] env[65385]: DEBUG nova.network.neutron [-] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 831.441428] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.445047] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.445047] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.505967] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.511238] env[65385]: WARNING neutronclient.v2_0.client [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.512473] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.512473] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.604357] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673525} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.604972] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b1cca536-e1d9-43f8-8a5c-be3981c106e7/b1cca536-e1d9-43f8-8a5c-be3981c106e7.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.605213] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.605878] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a0baed35-9e4a-43fa-844b-2a5f5f031e00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.615637] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453764, 'name': CloneVM_Task} progress is 93%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.617365] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 831.617365] env[65385]: value = "task-4453765" [ 831.617365] env[65385]: _type = "Task" [ 831.617365] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.631379] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453765, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.732110] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.732612] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.788991] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d63371-c356-450b-9a58-d2ea280cd407 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.802282] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1499fce8-0828-4664-8e36-b17d3026ddd3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.835938] env[65385]: WARNING neutronclient.v2_0.client [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.836633] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.836978] env[65385]: WARNING openstack [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.845245] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea9bec8-a402-4379-a3be-d9d2defc6a0d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.860142] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88723500-1fdf-4c82-9341-fcf90066edde {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.866581] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.878807] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.935762] env[65385]: DEBUG nova.network.neutron [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updated VIF entry in instance network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 831.936196] env[65385]: DEBUG nova.network.neutron [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updating instance_info_cache with network_info: [{"id": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "address": "fa:16:3e:01:88:e2", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf61ecab0-44", "ovs_interfaceid": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.101193] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453764, 'name': CloneVM_Task} progress is 93%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.105713] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "ad29afd9-3e52-4fac-977f-91774acad7d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.126944] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453765, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075157} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.127196] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.127924] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07013368-6332-4cde-8ae4-8f3cdbd967be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.150724] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] b1cca536-e1d9-43f8-8a5c-be3981c106e7/b1cca536-e1d9-43f8-8a5c-be3981c106e7.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.151055] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-083422d2-534e-4693-bf39-481d4672a550 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.171592] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 832.171592] env[65385]: value = "task-4453766" [ 832.171592] env[65385]: _type = "Task" [ 832.171592] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.182089] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453766, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.257450] env[65385]: DEBUG nova.compute.manager [req-e701c85e-24a3-4ee7-be65-b0cadb026b44 req-f786a0b0-8b06-48d2-8ff5-0ba2cf63846b service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Received event network-vif-deleted-29fa73f0-74dc-443a-b870-f4683fdc5d74 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 832.257573] env[65385]: INFO nova.compute.manager [req-e701c85e-24a3-4ee7-be65-b0cadb026b44 req-f786a0b0-8b06-48d2-8ff5-0ba2cf63846b service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Neutron deleted interface 29fa73f0-74dc-443a-b870-f4683fdc5d74; detaching it from the instance and deleting it from the info cache [ 832.257761] env[65385]: DEBUG nova.network.neutron [req-e701c85e-24a3-4ee7-be65-b0cadb026b44 req-f786a0b0-8b06-48d2-8ff5-0ba2cf63846b service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.301955] env[65385]: DEBUG nova.network.neutron [-] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.382637] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 832.441919] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] Releasing lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.441919] env[65385]: DEBUG nova.compute.manager [req-6b1a3b2b-d976-4dc0-a50a-f2e2183daf7d req-236244fb-4a1c-437c-9617-6f0ce8334bc5 service nova] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Received event network-vif-deleted-fe815bb4-491a-42f5-a5d3-2eba8997444a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 832.441919] env[65385]: DEBUG oslo_concurrency.lockutils [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Acquired lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 832.441919] env[65385]: DEBUG nova.network.neutron [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Refreshing network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 832.603659] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453764, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.683058] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453766, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.762459] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b910482-2dfc-40fb-9324-33bf497121d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.777528] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d99f32-aad2-4e78-aa81-6ce3d926811f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.808315] env[65385]: INFO nova.compute.manager [-] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Took 1.37 seconds to deallocate network for instance. [ 832.832338] env[65385]: DEBUG nova.compute.manager [req-e701c85e-24a3-4ee7-be65-b0cadb026b44 req-f786a0b0-8b06-48d2-8ff5-0ba2cf63846b service nova] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Detach interface failed, port_id=29fa73f0-74dc-443a-b870-f4683fdc5d74, reason: Instance 0673a0d2-76ac-4cd5-8cc3-b8596877c641 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 832.891759] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 832.892889] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.994s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 832.892889] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 26.556s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.895192] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 832.895312] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Cleaning up deleted instances {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11916}} [ 832.944047] env[65385]: WARNING neutronclient.v2_0.client [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 832.944801] env[65385]: WARNING openstack [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.945162] env[65385]: WARNING openstack [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.104165] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453764, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.129059] env[65385]: WARNING openstack [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.129059] env[65385]: WARNING openstack [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.185632] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453766, 'name': ReconfigVM_Task, 'duration_secs': 0.830228} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.185632] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Reconfigured VM instance instance-0000003d to attach disk [datastore2] b1cca536-e1d9-43f8-8a5c-be3981c106e7/b1cca536-e1d9-43f8-8a5c-be3981c106e7.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.186115] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d84183e-ef67-47f6-aac8-5a0d34567252 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.195336] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 833.195336] env[65385]: value = "task-4453767" [ 833.195336] env[65385]: _type = "Task" [ 833.195336] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.198142] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "f25cc41e-560e-4cab-8bf0-dac47899d037" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.198405] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "f25cc41e-560e-4cab-8bf0-dac47899d037" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.198612] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "f25cc41e-560e-4cab-8bf0-dac47899d037-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.198784] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "f25cc41e-560e-4cab-8bf0-dac47899d037-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.198946] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "f25cc41e-560e-4cab-8bf0-dac47899d037-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.202578] env[65385]: INFO nova.compute.manager [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Terminating instance [ 833.207136] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453767, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.213125] env[65385]: WARNING neutronclient.v2_0.client [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.214444] env[65385]: WARNING openstack [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.215207] env[65385]: WARNING openstack [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.333923] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.351807] env[65385]: DEBUG nova.network.neutron [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updated VIF entry in instance network info cache for port f61ecab0-4449-47db-8bfb-6ff1499f6d71. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 833.352363] env[65385]: DEBUG nova.network.neutron [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updating instance_info_cache with network_info: [{"id": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "address": "fa:16:3e:01:88:e2", "network": {"id": "0a9011dc-0679-44a3-940c-83538f5a6b1a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-638498222-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a3e5ed6b95f4282b1ef29d17fa58975", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf61ecab0-44", "ovs_interfaceid": "f61ecab0-4449-47db-8bfb-6ff1499f6d71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 833.398344] env[65385]: DEBUG nova.objects.instance [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lazy-loading 'migration_context' on Instance uuid 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.407418] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] There are 32 instances to clean {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11925}} [ 833.407700] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: e3248f4a-8985-48b0-84cf-795f00ff21fa] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 833.422653] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.422826] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.422930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.423139] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.423295] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.426660] env[65385]: INFO nova.compute.manager [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Terminating instance [ 833.603700] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453764, 'name': CloneVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.705864] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453767, 'name': Rename_Task, 'duration_secs': 0.172681} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.705864] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.706269] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d563fe0-5b1f-4ebb-ae66-baf53d0726ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.711877] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "refresh_cache-f25cc41e-560e-4cab-8bf0-dac47899d037" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.712115] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "refresh_cache-f25cc41e-560e-4cab-8bf0-dac47899d037" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.712357] env[65385]: DEBUG nova.network.neutron [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 833.717199] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 833.717199] env[65385]: value = "task-4453768" [ 833.717199] env[65385]: _type = "Task" [ 833.717199] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.726101] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453768, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.733723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "f521e37d-aa86-4533-9fbe-7d040ed90e53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.733723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.733723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "f521e37d-aa86-4533-9fbe-7d040ed90e53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.733723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.733723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.736482] env[65385]: INFO nova.compute.manager [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Terminating instance [ 833.856909] env[65385]: DEBUG oslo_concurrency.lockutils [req-2aa70410-ed3e-43de-89f7-2b586740de4c req-248b4357-c82a-4bc1-90d1-a76e14cfe1a5 service nova] Releasing lock "refresh_cache-7fcbdc79-688c-479f-94e2-f4542abe714b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.913945] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 4232f3a4-c42b-4b9c-8e1d-da591634ffbb] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 833.935150] env[65385]: DEBUG nova.compute.manager [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 833.935150] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.935150] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4183b5ce-bfc3-4377-ae74-bbbe346944f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.945459] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.945610] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5b0c282-a85e-4ad8-82d0-4098b884023c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.957020] env[65385]: DEBUG oslo_vmware.api [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 833.957020] env[65385]: value = "task-4453769" [ 833.957020] env[65385]: _type = "Task" [ 833.957020] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.970645] env[65385]: DEBUG oslo_vmware.api [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.109413] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453764, 'name': CloneVM_Task, 'duration_secs': 2.627457} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.109713] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Created linked-clone VM from snapshot [ 834.110575] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a00a4e-8548-4cc7-a626-e5224470f688 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.122868] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Uploading image 4f147f31-56c2-4e3c-b88b-8c468df0cf00 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 834.152694] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 834.152694] env[65385]: value = "vm-871058" [ 834.152694] env[65385]: _type = "VirtualMachine" [ 834.152694] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 834.153078] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-959b512a-c491-4ed3-aea7-b03cc64a14b3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.164230] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lease: (returnval){ [ 834.164230] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5215dbbc-2fc6-684b-d7ac-295f0ba7097f" [ 834.164230] env[65385]: _type = "HttpNfcLease" [ 834.164230] env[65385]: } obtained for exporting VM: (result){ [ 834.164230] env[65385]: value = "vm-871058" [ 834.164230] env[65385]: _type = "VirtualMachine" [ 834.164230] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 834.164570] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the lease: (returnval){ [ 834.164570] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5215dbbc-2fc6-684b-d7ac-295f0ba7097f" [ 834.164570] env[65385]: _type = "HttpNfcLease" [ 834.164570] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 834.173502] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 834.173502] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5215dbbc-2fc6-684b-d7ac-295f0ba7097f" [ 834.173502] env[65385]: _type = "HttpNfcLease" [ 834.173502] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 834.218287] env[65385]: WARNING neutronclient.v2_0.client [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 834.218974] env[65385]: WARNING openstack [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 834.219329] env[65385]: WARNING openstack [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 834.238728] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453768, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.241541] env[65385]: DEBUG nova.compute.manager [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 834.241751] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.243035] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735e4695-d01a-401e-acf9-9386f3c96749 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.247814] env[65385]: DEBUG nova.network.neutron [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 834.257255] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.257670] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0339949-2892-4a67-a49f-337acdbfb3e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.270508] env[65385]: DEBUG oslo_vmware.api [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 834.270508] env[65385]: value = "task-4453771" [ 834.270508] env[65385]: _type = "Task" [ 834.270508] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.281596] env[65385]: DEBUG oslo_vmware.api [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.328946] env[65385]: DEBUG nova.network.neutron [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 834.417876] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: fae095a8-a768-4263-b456-ed0745398755] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 834.475083] env[65385]: DEBUG oslo_vmware.api [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453769, 'name': PowerOffVM_Task, 'duration_secs': 0.240076} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.475083] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 834.475083] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 834.475083] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-829cf83b-762d-4eea-8921-86caa686733a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.544341] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8732ec-76c0-4f0c-9a54-4b4c30a54e8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.553241] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4cc004-7382-4c1d-977e-196cff028020 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.591759] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548eb8a7-1c07-4bb8-88c3-1772f8f03873 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.601291] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb72ec0b-d02a-4c7c-aa1d-f6ae278b4b5a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.619291] env[65385]: DEBUG nova.compute.provider_tree [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.675461] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 834.675461] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5215dbbc-2fc6-684b-d7ac-295f0ba7097f" [ 834.675461] env[65385]: _type = "HttpNfcLease" [ 834.675461] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 834.675933] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 834.675933] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5215dbbc-2fc6-684b-d7ac-295f0ba7097f" [ 834.675933] env[65385]: _type = "HttpNfcLease" [ 834.675933] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 834.676558] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65145713-9075-4595-9162-30fb0eb73c31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.685209] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52098147-0449-ab53-aa6b-da89958232e3/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 834.685446] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52098147-0449-ab53-aa6b-da89958232e3/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 834.753865] env[65385]: DEBUG oslo_vmware.api [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453768, 'name': PowerOnVM_Task, 'duration_secs': 0.696409} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.754151] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 834.754328] env[65385]: INFO nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Took 9.63 seconds to spawn the instance on the hypervisor. [ 834.754541] env[65385]: DEBUG nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 834.755575] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911935cf-f32c-49c6-89d2-d1a2e4dc3be8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.784298] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e852cce8-935b-455b-ba91-9a5bd845b2fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.785915] env[65385]: DEBUG oslo_vmware.api [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453771, 'name': PowerOffVM_Task, 'duration_secs': 0.430945} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.786265] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 834.786433] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 834.787808] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1ac6444-ed5d-432b-af04-a7636098579a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.835431] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "refresh_cache-f25cc41e-560e-4cab-8bf0-dac47899d037" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.836221] env[65385]: DEBUG nova.compute.manager [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 834.836456] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.838801] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a467969b-ab78-4628-a813-5bf24463c19a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.848049] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.848191] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ee2ad9c-a1a7-460f-af97-50d4f2ac6d70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.859202] env[65385]: DEBUG oslo_vmware.api [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 834.859202] env[65385]: value = "task-4453774" [ 834.859202] env[65385]: _type = "Task" [ 834.859202] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.868389] env[65385]: DEBUG oslo_vmware.api [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453774, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.921908] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 53c31e42-4db0-4428-8336-b3a82e3f0e8f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 835.126617] env[65385]: DEBUG nova.scheduler.client.report [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 835.277428] env[65385]: INFO nova.compute.manager [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Took 46.08 seconds to build instance. [ 835.370556] env[65385]: DEBUG oslo_vmware.api [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453774, 'name': PowerOffVM_Task, 'duration_secs': 0.241526} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.371064] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.371254] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.371565] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-271222cf-b5c2-4327-9d90-d029ecbfa808 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.401843] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.402122] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.402393] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleting the datastore file [datastore2] f25cc41e-560e-4cab-8bf0-dac47899d037 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.402672] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-687f4653-9125-415c-9276-e0c31100cb85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.414415] env[65385]: DEBUG oslo_vmware.api [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 835.414415] env[65385]: value = "task-4453776" [ 835.414415] env[65385]: _type = "Task" [ 835.414415] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.425858] env[65385]: DEBUG oslo_vmware.api [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453776, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.427869] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 34730159-3bec-4ad5-b85e-0f67998d6001] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 835.643636] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "0b02fcd5-9fc7-4543-a754-d31fa0186981" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.644060] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.780346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c4668e3e-ef46-411d-b050-d9915a12083a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.809s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.804380] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.804600] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.804790] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleting the datastore file [datastore1] b702ba27-bbc7-4030-a96d-30e1ed656a9b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.805098] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b7004f8-f6d8-4b93-8ef4-ee42c7594ef7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.815922] env[65385]: DEBUG oslo_vmware.api [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 835.815922] env[65385]: value = "task-4453777" [ 835.815922] env[65385]: _type = "Task" [ 835.815922] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.821763] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.822125] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.822379] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleting the datastore file [datastore1] f521e37d-aa86-4533-9fbe-7d040ed90e53 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.828187] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85ddb5a1-f027-45f4-afa2-f575ec0a77ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.833368] env[65385]: DEBUG oslo_vmware.api [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453777, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.835978] env[65385]: DEBUG oslo_vmware.api [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for the task: (returnval){ [ 835.835978] env[65385]: value = "task-4453778" [ 835.835978] env[65385]: _type = "Task" [ 835.835978] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.846028] env[65385]: DEBUG oslo_vmware.api [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.925603] env[65385]: DEBUG oslo_vmware.api [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453776, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162609} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.925854] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.926012] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 835.926192] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 835.926367] env[65385]: INFO nova.compute.manager [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Took 1.09 seconds to destroy the instance on the hypervisor. [ 835.926598] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 835.926792] env[65385]: DEBUG nova.compute.manager [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 835.926888] env[65385]: DEBUG nova.network.neutron [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 835.927161] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.927688] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.927942] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.935979] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: c798006c-3e74-461f-8397-3d8ca0be0a95] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 835.953665] env[65385]: DEBUG nova.network.neutron [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 835.953665] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.145557] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.253s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.154385] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.975s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 836.155822] env[65385]: INFO nova.compute.claims [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.160400] env[65385]: DEBUG nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 836.327300] env[65385]: DEBUG oslo_vmware.api [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453777, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215266} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.327581] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.327759] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.327948] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.328164] env[65385]: INFO nova.compute.manager [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Took 2.39 seconds to destroy the instance on the hypervisor. [ 836.328401] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 836.328605] env[65385]: DEBUG nova.compute.manager [-] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 836.328756] env[65385]: DEBUG nova.network.neutron [-] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 836.328961] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.329795] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.330815] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.347428] env[65385]: DEBUG oslo_vmware.api [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Task: {'id': task-4453778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209701} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.349019] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.349019] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.349019] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.349019] env[65385]: INFO nova.compute.manager [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Took 2.11 seconds to destroy the instance on the hypervisor. [ 836.349019] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 836.349286] env[65385]: DEBUG nova.compute.manager [-] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 836.349326] env[65385]: DEBUG nova.network.neutron [-] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 836.349589] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.350134] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.350432] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.415209] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.441573] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 2f66a03b-874b-4bee-9694-49813ecb0c8a] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 836.455817] env[65385]: DEBUG nova.network.neutron [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.467210] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.481798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.482041] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 836.482235] env[65385]: INFO nova.compute.manager [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Shelving [ 836.695855] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.851472] env[65385]: DEBUG nova.compute.manager [req-ba8d40c1-4d31-43ca-bd01-9ef08e8171a1 req-13014c53-0b18-4c54-bde2-13efe009cafc service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Received event network-vif-deleted-c897de6b-a79c-436e-af74-6db933cc0c9f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 836.851727] env[65385]: INFO nova.compute.manager [req-ba8d40c1-4d31-43ca-bd01-9ef08e8171a1 req-13014c53-0b18-4c54-bde2-13efe009cafc service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Neutron deleted interface c897de6b-a79c-436e-af74-6db933cc0c9f; detaching it from the instance and deleting it from the info cache [ 836.851845] env[65385]: DEBUG nova.network.neutron [req-ba8d40c1-4d31-43ca-bd01-9ef08e8171a1 req-13014c53-0b18-4c54-bde2-13efe009cafc service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.929899] env[65385]: DEBUG nova.compute.manager [req-5f861592-bb9b-46eb-a38e-d3d42ab74586 req-c049786c-6e8f-4eb2-93fd-f906c70ef784 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Received event network-vif-deleted-2034d6c4-ac55-40da-9bbe-a912b04d6548 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 836.930458] env[65385]: INFO nova.compute.manager [req-5f861592-bb9b-46eb-a38e-d3d42ab74586 req-c049786c-6e8f-4eb2-93fd-f906c70ef784 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Neutron deleted interface 2034d6c4-ac55-40da-9bbe-a912b04d6548; detaching it from the instance and deleting it from the info cache [ 836.930811] env[65385]: DEBUG nova.network.neutron [req-5f861592-bb9b-46eb-a38e-d3d42ab74586 req-c049786c-6e8f-4eb2-93fd-f906c70ef784 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.949516] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 28fd3828-38ec-426f-b4dc-1e094304780b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 836.961686] env[65385]: INFO nova.compute.manager [-] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Took 1.03 seconds to deallocate network for instance. [ 837.177368] env[65385]: DEBUG nova.network.neutron [-] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.222183] env[65385]: DEBUG nova.network.neutron [-] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.357215] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06bf813e-12d7-4cef-861d-43df6b27ef9f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.373170] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ebb08e-61ca-477d-be18-2cf69aafb3d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.420347] env[65385]: DEBUG nova.compute.manager [req-ba8d40c1-4d31-43ca-bd01-9ef08e8171a1 req-13014c53-0b18-4c54-bde2-13efe009cafc service nova] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Detach interface failed, port_id=c897de6b-a79c-436e-af74-6db933cc0c9f, reason: Instance b702ba27-bbc7-4030-a96d-30e1ed656a9b could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 837.437840] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e9611ad-b547-4593-80bb-d90cb8a24807 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.448528] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def739fb-a845-416c-8dab-5d54e943e79d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.465466] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 54ae4cd5-76d2-4e76-9528-6f511c4dca23] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 837.471990] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.497668] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 837.498057] env[65385]: DEBUG nova.compute.manager [req-5f861592-bb9b-46eb-a38e-d3d42ab74586 req-c049786c-6e8f-4eb2-93fd-f906c70ef784 service nova] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Detach interface failed, port_id=2034d6c4-ac55-40da-9bbe-a912b04d6548, reason: Instance f521e37d-aa86-4533-9fbe-7d040ed90e53 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 837.501955] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67fcf9dd-1775-4612-b6ce-f0beace21139 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.510780] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 837.510780] env[65385]: value = "task-4453779" [ 837.510780] env[65385]: _type = "Task" [ 837.510780] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.524509] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453779, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.687229] env[65385]: INFO nova.compute.manager [-] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Took 1.36 seconds to deallocate network for instance. [ 837.703560] env[65385]: INFO nova.compute.manager [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Swapping old allocation on dict_keys(['1af23b69-5ce6-4d6c-8591-1b95ecca8a6b']) held by migration 40af7784-05e8-4465-ba0d-2e13af06eb72 for instance [ 837.728191] env[65385]: INFO nova.compute.manager [-] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Took 1.38 seconds to deallocate network for instance. [ 837.745343] env[65385]: DEBUG nova.scheduler.client.report [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Overwriting current allocation {'allocations': {'1af23b69-5ce6-4d6c-8591-1b95ecca8a6b': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 88}}, 'project_id': '99173b1f52aa42f9b0db92366db8c411', 'user_id': 'ae527c3a64084ef388ff9415f1a6f883', 'consumer_generation': 1} on consumer 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22 {{(pid=65385) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 837.811537] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 837.879748] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94019059-f6bc-4cae-bf98-21417533953b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.884760] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.884934] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquired lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 837.885119] env[65385]: DEBUG nova.network.neutron [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 837.889253] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a7a64f-19e9-415b-a2f6-f5b4283e2fa6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.922687] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c05ee0-8d3f-4264-a5e9-5209e2b13aa9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.931870] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e39780-afce-4e5b-af6a-d15ef2ec8c93 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.949667] env[65385]: DEBUG nova.compute.provider_tree [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.969382] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 8b478626-e3dc-42c4-bd59-fa0a71d798c9] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 838.021144] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453779, 'name': PowerOffVM_Task, 'duration_secs': 0.208984} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.021144] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 838.021858] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5504ff98-ea71-4dc3-abce-9d8d23820877 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.042017] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311d3e91-93e7-4d26-a33d-b9237e5b9df4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.202183] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.236020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.389701] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.390851] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.391454] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.454048] env[65385]: DEBUG nova.scheduler.client.report [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 838.474556] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 2e4c064f-27b0-4ccb-8fe6-c34a61153663] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 838.519303] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.520318] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.554633] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 838.555515] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-be4b46ec-f7b0-42de-9f89-0f37824bdc08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.570566] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 838.570566] env[65385]: value = "task-4453780" [ 838.570566] env[65385]: _type = "Task" [ 838.570566] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.586993] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453780, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.594569] env[65385]: WARNING neutronclient.v2_0.client [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.595273] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.595651] env[65385]: WARNING openstack [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.684530] env[65385]: DEBUG nova.network.neutron [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance_info_cache with network_info: [{"id": "ad95d76f-e023-47b2-a936-e7ed11779276", "address": "fa:16:3e:b0:c7:d2", "network": {"id": "49354c6d-7b52-4e9f-ab59-89bac4fb36b8", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "51b61584e66c48498e8f79f7febc4707", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95d76f-e0", "ovs_interfaceid": "ad95d76f-e023-47b2-a936-e7ed11779276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 838.960212] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.806s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.961053] env[65385]: DEBUG nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 838.965198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.287s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.965489] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.968480] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.838s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.968749] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.971486] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.439s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.971752] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.974257] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.978s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.974532] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.977298] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 24.544s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.981089] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 950efcc6-7e4f-4b30-b0fa-d940f893e1d5] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 839.017896] env[65385]: INFO nova.scheduler.client.report [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Deleted allocations for instance e9fd0c9f-5795-44df-9dcb-982b7722ae69 [ 839.020839] env[65385]: INFO nova.scheduler.client.report [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Deleted allocations for instance 0115dde8-b0aa-4960-a452-ced6587bc567 [ 839.034157] env[65385]: INFO nova.scheduler.client.report [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleted allocations for instance 98430d7e-626e-4532-802e-38f7fc583ae3 [ 839.041037] env[65385]: INFO nova.scheduler.client.report [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Deleted allocations for instance ab8df1d8-615d-4083-bdf4-25f679a64982 [ 839.084994] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453780, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.187957] env[65385]: DEBUG oslo_concurrency.lockutils [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Releasing lock "refresh_cache-38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.188530] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.188853] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ad15e20-ba75-4591-b804-528ec4bb447b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.197107] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 839.197107] env[65385]: value = "task-4453781" [ 839.197107] env[65385]: _type = "Task" [ 839.197107] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.206815] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.482836] env[65385]: INFO nova.compute.claims [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.487999] env[65385]: DEBUG nova.compute.utils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 839.489765] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: a71d2fa6-006c-4f76-8ae8-467e149297da] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 839.492036] env[65385]: DEBUG nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 839.492206] env[65385]: DEBUG nova.network.neutron [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 839.492613] env[65385]: WARNING neutronclient.v2_0.client [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 839.493045] env[65385]: WARNING neutronclient.v2_0.client [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 839.494174] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.494531] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.533670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae5e6c99-40ea-41d2-bc19-ec1544030c44 tempest-ServersTestMultiNic-534525242 tempest-ServersTestMultiNic-534525242-project-member] Lock "e9fd0c9f-5795-44df-9dcb-982b7722ae69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.738s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.534731] env[65385]: DEBUG oslo_concurrency.lockutils [None req-87f6c89a-7e0a-4167-aa9d-1794b6c7c2bc tempest-SecurityGroupsTestJSON-1673006647 tempest-SecurityGroupsTestJSON-1673006647-project-member] Lock "0115dde8-b0aa-4960-a452-ced6587bc567" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.217s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.547919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eaded57e-db1f-4856-89c8-9093f1d5e7e3 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "98430d7e-626e-4532-802e-38f7fc583ae3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.784s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.552639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ae2d9f0-59a6-4e60-ad21-e3ec7d00c4c9 tempest-ServerAddressesTestJSON-84004151 tempest-ServerAddressesTestJSON-84004151-project-member] Lock "ab8df1d8-615d-4083-bdf4-25f679a64982" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.095s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.557286] env[65385]: DEBUG nova.policy [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 839.585099] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453780, 'name': CreateSnapshot_Task, 'duration_secs': 0.712676} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.585099] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 839.585733] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9436537a-3dcd-43d7-9309-e16424e270f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.709954] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453781, 'name': PowerOffVM_Task, 'duration_secs': 0.189408} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.711540] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.711540] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:48:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f35b2975-b7d3-4f0c-9c41-d7d5c82d8413',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1230048441',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 839.711540] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 839.711872] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 839.711872] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 839.711872] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 839.711995] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 839.712217] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.712369] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 839.712584] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 839.712748] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 839.712918] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 839.718927] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2eb23504-5154-4c96-99ba-06a0c5d64eb5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.738056] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 839.738056] env[65385]: value = "task-4453782" [ 839.738056] env[65385]: _type = "Task" [ 839.738056] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.747729] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453782, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.910814] env[65385]: DEBUG nova.network.neutron [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Successfully created port: 7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 839.996233] env[65385]: INFO nova.compute.resource_tracker [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating resource usage from migration 73856ca4-6bf7-447d-9523-fd6a10d8cb2a [ 840.000713] env[65385]: DEBUG nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 840.006201] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5291c04c-24ad-4c64-9fc2-fde42da4bc0f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 840.114612] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 840.118194] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-01486033-0bce-4282-b3a4-915fd7b97131 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.129679] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 840.129679] env[65385]: value = "task-4453783" [ 840.129679] env[65385]: _type = "Task" [ 840.129679] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.141673] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453783, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.253667] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453782, 'name': ReconfigVM_Task, 'duration_secs': 0.182128} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.253667] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b1e6c6-26a9-400a-8d22-14946586c787 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.119108] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: de8632e0-bf38-440d-b6a3-895efeef122d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 841.119423] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:48:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f35b2975-b7d3-4f0c-9c41-d7d5c82d8413',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1230048441',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 841.119423] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 841.119535] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 841.120163] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 841.120163] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 841.120163] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 841.120163] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.120386] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 841.120480] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 841.120595] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 841.120757] env[65385]: DEBUG nova.virt.hardware [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 841.127510] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf5505f2-3885-486d-a999-43bbaf8cdc8d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.139101] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453783, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.139101] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 841.139101] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5230b05c-928d-c9b2-5c6d-c724a5d698c0" [ 841.139101] env[65385]: _type = "Task" [ 841.139101] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.148586] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5230b05c-928d-c9b2-5c6d-c724a5d698c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.425886] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131c41dc-7197-4ba6-9f63-aa99196b5249 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.438531] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c9d9d2-d31f-4eab-a5b9-ac10e9d09d00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.478565] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43c23ae-ddb9-44b9-99a6-40278b79a4ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.486750] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d219bbe6-a6f5-4a81-b272-e75d63fce1cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.502070] env[65385]: DEBUG nova.compute.provider_tree [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.554890] env[65385]: DEBUG nova.network.neutron [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Successfully updated port: 7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 841.625514] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 69fdd5df-a9f2-486f-8a79-87c034366083] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 841.629293] env[65385]: DEBUG nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 841.637920] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453783, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.650690] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5230b05c-928d-c9b2-5c6d-c724a5d698c0, 'name': SearchDatastore_Task, 'duration_secs': 0.014057} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.657389] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfiguring VM instance instance-00000029 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 841.660423] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0376f017-d199-457e-8f07-26e2521f2d2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.682220] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 841.683151] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 841.683151] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 841.683151] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 841.683151] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 841.683151] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 841.683360] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.683588] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 841.683760] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 841.683918] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 841.684108] env[65385]: DEBUG nova.virt.hardware [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 841.685437] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e79fc1-763b-4812-998e-d3741d884532 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.689705] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 841.689705] env[65385]: value = "task-4453784" [ 841.689705] env[65385]: _type = "Task" [ 841.689705] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.698242] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb3d750-0cf7-47e7-a029-f6fda4c69244 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.706362] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453784, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.005598] env[65385]: DEBUG nova.scheduler.client.report [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 842.057651] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.057946] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 842.058155] env[65385]: DEBUG nova.network.neutron [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 842.133336] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453783, 'name': CloneVM_Task, 'duration_secs': 1.697311} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.133336] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Created linked-clone VM from snapshot [ 842.133825] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ff024f-0559-4d0f-bc04-509de6fe2e7d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.137215] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 427d5c6d-ab01-42ef-8ee5-edd608896b8b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 842.146095] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Uploading image 68865751-553d-4087-a64b-e593c2630167 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 842.191335] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 842.191335] env[65385]: value = "vm-871060" [ 842.191335] env[65385]: _type = "VirtualMachine" [ 842.191335] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 842.191947] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7283d98f-c1a6-4e1a-abc9-a3d53dde02af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.204962] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453784, 'name': ReconfigVM_Task, 'duration_secs': 0.199941} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.206971] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfigured VM instance instance-00000029 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 842.207488] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lease: (returnval){ [ 842.207488] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc9c3b-805f-6d95-55ea-dfdba515c3b9" [ 842.207488] env[65385]: _type = "HttpNfcLease" [ 842.207488] env[65385]: } obtained for exporting VM: (result){ [ 842.207488] env[65385]: value = "vm-871060" [ 842.207488] env[65385]: _type = "VirtualMachine" [ 842.207488] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 842.207878] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the lease: (returnval){ [ 842.207878] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc9c3b-805f-6d95-55ea-dfdba515c3b9" [ 842.207878] env[65385]: _type = "HttpNfcLease" [ 842.207878] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 842.211404] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78228ca4-361d-41c1-948c-89f466c4f6bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.220889] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 842.220889] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc9c3b-805f-6d95-55ea-dfdba515c3b9" [ 842.220889] env[65385]: _type = "HttpNfcLease" [ 842.220889] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 842.242171] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.242584] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d00fd43-1632-453d-a5f8-ee880ba30817 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.264430] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 842.264430] env[65385]: value = "task-4453786" [ 842.264430] env[65385]: _type = "Task" [ 842.264430] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.276996] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453786, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.511152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.534s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.511470] env[65385]: INFO nova.compute.manager [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Migrating [ 842.518525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.228s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.518803] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.521110] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.210s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.521365] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.524804] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.423s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.525088] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.528062] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.124s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.528284] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.530275] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.294s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.532020] env[65385]: INFO nova.compute.claims [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.563036] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.563036] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.572121] env[65385]: INFO nova.scheduler.client.report [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Deleted allocations for instance 1ad1e02b-fb82-4a88-9047-be316d590bd2 [ 842.581033] env[65385]: INFO nova.scheduler.client.report [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Deleted allocations for instance 483cf486-f0fc-48a4-9db2-970d0f35865a [ 842.608765] env[65385]: INFO nova.scheduler.client.report [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Deleted allocations for instance 3657ec21-3ec7-44c9-92d1-570655b58f26 [ 842.615051] env[65385]: INFO nova.scheduler.client.report [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Deleted allocations for instance 59450963-d83f-46e8-8c13-05d4f1818c64 [ 842.641208] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 1bc94ee7-d4f9-48c3-97f1-9e662eb50582] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 842.721035] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 842.721035] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc9c3b-805f-6d95-55ea-dfdba515c3b9" [ 842.721035] env[65385]: _type = "HttpNfcLease" [ 842.721035] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 842.721510] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 842.721510] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc9c3b-805f-6d95-55ea-dfdba515c3b9" [ 842.721510] env[65385]: _type = "HttpNfcLease" [ 842.721510] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 842.722246] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1115b7e1-0def-4daa-a100-652f1fa1bd1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.734162] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52913248-27d5-9d74-16ad-44e60466743c/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 842.734926] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52913248-27d5-9d74-16ad-44e60466743c/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 842.799729] env[65385]: DEBUG nova.network.neutron [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 842.816200] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453786, 'name': ReconfigVM_Task, 'duration_secs': 0.300563} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.816200] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22/38e0ebf5-1d92-4906-b0bd-b0ebf8651f22.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.816200] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a8a714-0e2a-45ea-a53c-bac9dc1f90d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.841109] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52098147-0449-ab53-aa6b-da89958232e3/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 842.842036] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebee10e0-fe73-4e43-b738-fea64d40eb5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.845704] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ff5058-335c-4730-8b7d-329c8516974e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.848684] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9ac49d14-69c6-436d-9525-e94d2a0c987e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.854169] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52098147-0449-ab53-aa6b-da89958232e3/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 842.854231] env[65385]: ERROR oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52098147-0449-ab53-aa6b-da89958232e3/disk-0.vmdk due to incomplete transfer. [ 842.870180] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-60363582-bcaa-422e-a12a-1e319e12b072 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.874460] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e34c657-4174-47f5-ac9a-461fe64c3dd9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.897717] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1519f0-1f76-4e0e-8573-8e03c5ef4121 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.900555] env[65385]: DEBUG oslo_vmware.rw_handles [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52098147-0449-ab53-aa6b-da89958232e3/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 842.900756] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Uploaded image 4f147f31-56c2-4e3c-b88b-8c468df0cf00 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 842.903267] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 842.904213] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c92646cb-d302-4e4d-9174-ee862877806b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.911589] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.918221] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ee50841-5f92-4b47-a394-eac9e4ef8ce5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.920180] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 842.920180] env[65385]: value = "task-4453787" [ 842.920180] env[65385]: _type = "Task" [ 842.920180] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.926981] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 842.926981] env[65385]: value = "task-4453788" [ 842.926981] env[65385]: _type = "Task" [ 842.926981] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.934367] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453787, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.940488] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453788, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.977488] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.977488] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.049446] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.049690] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.049874] env[65385]: DEBUG nova.network.neutron [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 843.093784] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09891518-0f1f-4219-aa4b-a3f3262b6781 tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "483cf486-f0fc-48a4-9db2-970d0f35865a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.377s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.096404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2a6711c-02fb-4e2e-8323-e98ac3c71ad3 tempest-ServersTestBootFromVolume-1171683164 tempest-ServersTestBootFromVolume-1171683164-project-member] Lock "1ad1e02b-fb82-4a88-9047-be316d590bd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.477s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.120694] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d886ed25-53b6-4d4c-8dd1-5c898eed990f tempest-ServersAdminNegativeTestJSON-1084595625 tempest-ServersAdminNegativeTestJSON-1084595625-project-member] Lock "3657ec21-3ec7-44c9-92d1-570655b58f26" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.524s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.125856] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eb8631d9-5e33-4d91-805f-1b8a3023654e tempest-ServersTestFqdnHostnames-537130554 tempest-ServersTestFqdnHostnames-537130554-project-member] Lock "59450963-d83f-46e8-8c13-05d4f1818c64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.708s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.146076] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 60067529-1071-4295-b1c9-21523bf347d9] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 843.213017] env[65385]: DEBUG nova.compute.manager [req-db6fb492-43b1-4e10-b87e-e7281eee1dcc req-ea6d0596-677e-4a99-bd5b-d25a5a389338 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received event network-vif-plugged-7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 843.213342] env[65385]: DEBUG oslo_concurrency.lockutils [req-db6fb492-43b1-4e10-b87e-e7281eee1dcc req-ea6d0596-677e-4a99-bd5b-d25a5a389338 service nova] Acquiring lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.214260] env[65385]: DEBUG oslo_concurrency.lockutils [req-db6fb492-43b1-4e10-b87e-e7281eee1dcc req-ea6d0596-677e-4a99-bd5b-d25a5a389338 service nova] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.214620] env[65385]: DEBUG oslo_concurrency.lockutils [req-db6fb492-43b1-4e10-b87e-e7281eee1dcc req-ea6d0596-677e-4a99-bd5b-d25a5a389338 service nova] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.214808] env[65385]: DEBUG nova.compute.manager [req-db6fb492-43b1-4e10-b87e-e7281eee1dcc req-ea6d0596-677e-4a99-bd5b-d25a5a389338 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] No waiting events found dispatching network-vif-plugged-7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 843.214974] env[65385]: WARNING nova.compute.manager [req-db6fb492-43b1-4e10-b87e-e7281eee1dcc req-ea6d0596-677e-4a99-bd5b-d25a5a389338 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received unexpected event network-vif-plugged-7d96802a-1199-4a18-9a3d-f7b69a52590c for instance with vm_state building and task_state spawning. [ 843.437491] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453787, 'name': Destroy_Task, 'duration_secs': 0.445177} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.438011] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Destroyed the VM [ 843.438494] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 843.438619] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5d25a6e3-3cf3-4cb4-8865-decc2044651c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.445628] env[65385]: DEBUG oslo_vmware.api [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453788, 'name': PowerOnVM_Task, 'duration_secs': 0.451273} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.446357] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.453503] env[65385]: WARNING neutronclient.v2_0.client [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.453503] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.453599] env[65385]: WARNING openstack [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.466877] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 843.466877] env[65385]: value = "task-4453789" [ 843.466877] env[65385]: _type = "Task" [ 843.466877] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.480097] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453789, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.556670] env[65385]: WARNING neutronclient.v2_0.client [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.557432] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.557966] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.662044] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5df7d4d7-3bfb-4281-a3e4-8d40ddf3ef64] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 843.827971] env[65385]: DEBUG nova.network.neutron [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updating instance_info_cache with network_info: [{"id": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "address": "fa:16:3e:42:06:97", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d96802a-11", "ovs_interfaceid": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 843.978851] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453789, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.075540] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337af94a-29b1-4b25-9d7b-379c10990c32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.085071] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb9b1f4-4046-4b22-bc1a-e1dc06b1cc82 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.120510] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09641783-a5cd-492d-8e9a-c3c1512a7831 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.130231] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f38a28-fedb-4058-9af6-627b512d9a99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.145151] env[65385]: DEBUG nova.compute.provider_tree [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.165350] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: f77d861a-ae66-44c3-901d-e499926d6f5b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 844.298633] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.299251] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.331867] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 844.332356] env[65385]: DEBUG nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Instance network_info: |[{"id": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "address": "fa:16:3e:42:06:97", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d96802a-11", "ovs_interfaceid": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 844.332930] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:06:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d96802a-1199-4a18-9a3d-f7b69a52590c', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.342912] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Creating folder: Project (1991528775bd4557ae6150b30095e923). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.344593] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e85d7c9a-49c2-4ed8-9ac0-af80f72be309 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.358926] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Created folder: Project (1991528775bd4557ae6150b30095e923) in parent group-v870881. [ 844.359611] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Creating folder: Instances. Parent ref: group-v871061. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.360060] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5fa2554-5944-4de4-b47c-05486ee8aace {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.374431] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Created folder: Instances in parent group-v871061. [ 844.375321] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 844.375321] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.375516] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccf66c60-8548-4679-b1d9-0c5fee8dceb2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.401286] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.401286] env[65385]: value = "task-4453792" [ 844.401286] env[65385]: _type = "Task" [ 844.401286] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.410947] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453792, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.460508] env[65385]: INFO nova.compute.manager [None req-94489a19-4981-4978-acfb-9385867f0cad tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance to original state: 'active' [ 844.482944] env[65385]: DEBUG oslo_vmware.api [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453789, 'name': RemoveSnapshot_Task, 'duration_secs': 0.683794} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.483330] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 844.483594] env[65385]: INFO nova.compute.manager [None req-fc481249-9263-43c4-b473-0a61d79c3ebe tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Took 16.02 seconds to snapshot the instance on the hypervisor. [ 844.553376] env[65385]: WARNING neutronclient.v2_0.client [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 844.554059] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.554434] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.654394] env[65385]: DEBUG nova.scheduler.client.report [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 844.668795] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 649b4e21-fc55-415c-a1f3-ec724397b874] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 844.709304] env[65385]: DEBUG nova.network.neutron [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 844.912700] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453792, 'name': CreateVM_Task, 'duration_secs': 0.363544} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.912929] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.913490] env[65385]: WARNING neutronclient.v2_0.client [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 844.913959] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.914111] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 844.914335] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 844.914652] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-192068f2-0a86-4156-aa09-507f82d50844 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.920203] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 844.920203] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52149524-8fcb-6509-df5d-f7da4c0d1e03" [ 844.920203] env[65385]: _type = "Task" [ 844.920203] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.929742] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52149524-8fcb-6509-df5d-f7da4c0d1e03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.162021] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.629s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.162021] env[65385]: DEBUG nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 845.163250] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.495s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.165092] env[65385]: INFO nova.compute.claims [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.173510] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: c5b85574-df14-4a7a-ada0-c9668bf18a33] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 845.214491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 845.323921] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.324128] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.447220] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52149524-8fcb-6509-df5d-f7da4c0d1e03, 'name': SearchDatastore_Task, 'duration_secs': 0.011533} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.447220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 845.447220] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.447590] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.448458] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 845.448787] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.451570] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5de7f436-6922-4a2c-ba6c-7a26a6d1caad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.462411] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.463048] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.463650] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c90197f9-4b08-436c-b592-c7a1164638b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.470715] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 845.470715] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52236b2f-0f4d-a6b8-4b91-8d9b04142246" [ 845.470715] env[65385]: _type = "Task" [ 845.470715] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.479806] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52236b2f-0f4d-a6b8-4b91-8d9b04142246, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.670547] env[65385]: DEBUG nova.compute.utils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 845.675631] env[65385]: DEBUG nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 845.675631] env[65385]: DEBUG nova.network.neutron [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 845.675631] env[65385]: WARNING neutronclient.v2_0.client [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.675631] env[65385]: WARNING neutronclient.v2_0.client [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.676249] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.676729] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.684709] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 8ac44124-d8cf-421c-9ae4-c943df94550a] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 845.827613] env[65385]: DEBUG nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 845.985335] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52236b2f-0f4d-a6b8-4b91-8d9b04142246, 'name': SearchDatastore_Task, 'duration_secs': 0.010808} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.986097] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fd70f1c-9d0b-4a1f-ae06-eb756ad7b843 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.993419] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 845.993419] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526ef849-7724-a73c-0fe5-00276891a6f6" [ 845.993419] env[65385]: _type = "Task" [ 845.993419] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.002757] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526ef849-7724-a73c-0fe5-00276891a6f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.175455] env[65385]: DEBUG nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 846.189901] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: bf79fbcd-e943-4ff3-bcf8-6a2817df8292] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 846.357116] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 846.511469] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526ef849-7724-a73c-0fe5-00276891a6f6, 'name': SearchDatastore_Task, 'duration_secs': 0.012848} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.511873] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.512539] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] c18b5bde-1a17-4954-a2c5-582815ebfec8/c18b5bde-1a17-4954-a2c5-582815ebfec8.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.512918] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ccb2e373-e79e-4aaa-9847-039980b10a26 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.528052] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 846.528052] env[65385]: value = "task-4453793" [ 846.528052] env[65385]: _type = "Task" [ 846.528052] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.540133] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453793, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.697062] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: a8c76329-5056-4cf0-ac32-85d46429d3db] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 846.727931] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8802cfaf-8e8b-4dd2-86ac-52a816b40d85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.738222] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadac26b-4473-4a79-85ee-469cf8c3f707 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.752569] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bdb4ef-1526-411f-b2ac-b3c34f68ebea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.773767] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance '89f3c815-8671-47ce-9e74-bf6e652bb3c3' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 846.810588] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f6f755-bfc0-4c92-9bf9-541b27f46be3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.821032] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d52883c-0b52-40f7-a419-8ae5defd5e2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.838668] env[65385]: DEBUG nova.compute.provider_tree [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.896523] env[65385]: DEBUG nova.policy [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b26caf958b047c79ed5d7f3d4d6b883', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '647f4e3157c34a8f9a64cb8ba8b72da6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 847.038751] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453793, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.188975] env[65385]: DEBUG nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 847.206150] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: ef235d8e-6193-4010-ac50-72cd275e965a] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 847.220864] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 847.221140] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 847.221446] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 847.221533] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 847.221704] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 847.221774] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 847.221951] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.222121] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 847.222286] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 847.222459] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 847.222657] env[65385]: DEBUG nova.virt.hardware [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 847.223561] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c9bea9-f4cc-416d-b721-6c5435a072fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.234163] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62286a8-bbf0-49d2-a679-5330bae12b55 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.281331] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.281511] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fdcab2c-fd8c-4e6c-945e-87e8b1e50a1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.290157] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 847.290157] env[65385]: value = "task-4453794" [ 847.290157] env[65385]: _type = "Task" [ 847.290157] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.302543] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 847.302938] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance '89f3c815-8671-47ce-9e74-bf6e652bb3c3' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 847.346197] env[65385]: DEBUG nova.scheduler.client.report [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 847.541409] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453793, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516218} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.541644] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] c18b5bde-1a17-4954-a2c5-582815ebfec8/c18b5bde-1a17-4954-a2c5-582815ebfec8.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.541892] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.542221] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a648b512-9955-46d8-89b8-4bc5f42508dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.551496] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 847.551496] env[65385]: value = "task-4453795" [ 847.551496] env[65385]: _type = "Task" [ 847.551496] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.562500] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453795, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.715344] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 065659ae-2163-4d6b-a905-fc99d12b6790] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 847.727965] env[65385]: DEBUG nova.network.neutron [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Successfully created port: 97b8305b-a557-46c0-9f9a-f45d7ccd6e3a {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 847.811771] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 847.812020] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 847.812274] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 847.812546] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 847.812718] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 847.812922] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 847.813262] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.813413] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 847.813626] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 847.813798] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 847.814034] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 847.819896] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a535530-0d6e-49e8-8289-cafd000669f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.839683] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 847.839683] env[65385]: value = "task-4453796" [ 847.839683] env[65385]: _type = "Task" [ 847.839683] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.850360] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453796, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.852639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.689s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.854245] env[65385]: DEBUG nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 847.859202] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.676s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.859492] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.004s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.861717] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.272s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.867479] env[65385]: INFO nova.compute.claims [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.911880] env[65385]: INFO nova.scheduler.client.report [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted allocations for instance c5d00bb5-d869-466f-826e-61622e57959f [ 848.066973] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453795, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082891} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.067506] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.068618] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2007b4-f1b1-48ec-8a3a-23606d0512a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.107806] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] c18b5bde-1a17-4954-a2c5-582815ebfec8/c18b5bde-1a17-4954-a2c5-582815ebfec8.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.109225] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6d60fcd-48ea-42ff-b03c-16da62142a0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.134139] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 848.134139] env[65385]: value = "task-4453797" [ 848.134139] env[65385]: _type = "Task" [ 848.134139] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.146604] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453797, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.219942] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: a06f1bbf-b4f2-4ddb-81a8-1e10df8077d3] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 848.352781] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453796, 'name': ReconfigVM_Task, 'duration_secs': 0.16371} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.353195] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance '89f3c815-8671-47ce-9e74-bf6e652bb3c3' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 848.373642] env[65385]: DEBUG nova.compute.utils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 848.377881] env[65385]: DEBUG nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 848.378371] env[65385]: DEBUG nova.network.neutron [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 848.379180] env[65385]: WARNING neutronclient.v2_0.client [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 848.379640] env[65385]: WARNING neutronclient.v2_0.client [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 848.380493] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 848.380930] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 848.422538] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c2ed6c85-cfa4-4e11-8739-d6d93649d65a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "c5d00bb5-d869-466f-826e-61622e57959f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.692s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.645994] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453797, 'name': ReconfigVM_Task, 'duration_secs': 0.341837} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.646406] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Reconfigured VM instance instance-0000003e to attach disk [datastore1] c18b5bde-1a17-4954-a2c5-582815ebfec8/c18b5bde-1a17-4954-a2c5-582815ebfec8.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.647114] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50d7d1c0-9800-46db-9d7a-b8d9fbfaa58f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.657211] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 848.657211] env[65385]: value = "task-4453798" [ 848.657211] env[65385]: _type = "Task" [ 848.657211] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.667276] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453798, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.724953] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 6e1eeff2-1587-44c6-a154-f73a2d0f8061] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 848.861397] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 848.861843] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 848.861843] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 848.862172] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 848.862372] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 848.862753] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 848.862858] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.863082] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 848.863276] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 848.863440] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 848.863637] env[65385]: DEBUG nova.virt.hardware [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 848.869238] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Reconfiguring VM instance instance-00000019 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 848.869560] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27b1c958-3e82-4a66-bd1e-9a002324a2c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.884520] env[65385]: DEBUG nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 848.896503] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 848.896503] env[65385]: value = "task-4453799" [ 848.896503] env[65385]: _type = "Task" [ 848.896503] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.909710] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453799, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.170065] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453798, 'name': Rename_Task, 'duration_secs': 0.15847} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.170065] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.170065] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05538e44-d839-41b1-b9ac-60668f12c47c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.177423] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 849.177423] env[65385]: value = "task-4453800" [ 849.177423] env[65385]: _type = "Task" [ 849.177423] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.187294] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453800, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.227883] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 504f5660-6715-4c7a-965b-9d2ef1852391] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 849.414508] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453799, 'name': ReconfigVM_Task, 'duration_secs': 0.321722} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.415131] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Reconfigured VM instance instance-00000019 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 849.417281] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83e335e-c954-4b1a-b7a5-23fb5a7b2827 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.451793] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] 89f3c815-8671-47ce-9e74-bf6e652bb3c3/89f3c815-8671-47ce-9e74-bf6e652bb3c3.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.455382] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1217862-1907-49d6-9645-fa6d3b940bcd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.476927] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 849.476927] env[65385]: value = "task-4453801" [ 849.476927] env[65385]: _type = "Task" [ 849.476927] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.486795] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453801, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.489202] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170c8632-ea1f-440f-94ce-a0da433c3c28 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.497197] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b618ba4c-4fea-4841-8fa6-13463d546fa2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.541978] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da9a5d6-c75f-4db5-954a-341053789eec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.555091] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e1ba7f-87fc-4b9b-9bb2-ea6a7fc29e7b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.579463] env[65385]: DEBUG nova.compute.provider_tree [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.690022] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453800, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.699069] env[65385]: DEBUG nova.network.neutron [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Successfully updated port: 97b8305b-a557-46c0-9f9a-f45d7ccd6e3a {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 849.732086] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: b72885c3-5146-42a5-82e8-444cfd89413d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 849.746186] env[65385]: DEBUG nova.policy [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0f069272314a31ad33c77137513a17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c27eb6232a2148a1a259f57494b4ae30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 849.881205] env[65385]: DEBUG nova.compute.manager [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received event network-changed-7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 849.881205] env[65385]: DEBUG nova.compute.manager [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Refreshing instance network info cache due to event network-changed-7d96802a-1199-4a18-9a3d-f7b69a52590c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 849.881205] env[65385]: DEBUG oslo_concurrency.lockutils [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Acquiring lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.881205] env[65385]: DEBUG oslo_concurrency.lockutils [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Acquired lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.881503] env[65385]: DEBUG nova.network.neutron [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Refreshing network info cache for port 7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 849.905825] env[65385]: DEBUG nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 849.943992] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 849.945052] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 849.945052] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 849.945052] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 849.945052] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 849.945052] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 849.945609] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.945863] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 849.946225] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 849.947029] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 849.947029] env[65385]: DEBUG nova.virt.hardware [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 849.947706] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2380fef1-9467-4a10-a07e-003e12cf3c7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.957655] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4db6e8-3f63-492a-8a1b-9084365afd1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.989022] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453801, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.084442] env[65385]: DEBUG nova.scheduler.client.report [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 850.193591] env[65385]: DEBUG oslo_vmware.api [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453800, 'name': PowerOnVM_Task, 'duration_secs': 0.646679} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.194457] env[65385]: DEBUG nova.network.neutron [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Successfully created port: d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 850.197612] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.197920] env[65385]: INFO nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Took 8.57 seconds to spawn the instance on the hypervisor. [ 850.198214] env[65385]: DEBUG nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 850.199388] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915f5323-1eb0-4651-8e6d-9bd62a02fee5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.203522] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "refresh_cache-703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.203917] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "refresh_cache-703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 850.204164] env[65385]: DEBUG nova.network.neutron [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 850.240046] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 850.240299] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Cleaning up deleted instances with incomplete migration {{(pid=65385) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11954}} [ 850.390422] env[65385]: WARNING neutronclient.v2_0.client [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.391797] env[65385]: WARNING openstack [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.391797] env[65385]: WARNING openstack [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.421703] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "90c36f67-0c25-4487-a5e4-89827290953b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.422044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "90c36f67-0c25-4487-a5e4-89827290953b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.422293] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "90c36f67-0c25-4487-a5e4-89827290953b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.422498] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "90c36f67-0c25-4487-a5e4-89827290953b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.422708] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "90c36f67-0c25-4487-a5e4-89827290953b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.425309] env[65385]: INFO nova.compute.manager [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Terminating instance [ 850.491858] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453801, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.592793] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.731s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.593416] env[65385]: DEBUG nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 850.597555] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.120s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.597983] env[65385]: DEBUG nova.objects.instance [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 850.662532] env[65385]: WARNING openstack [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.662922] env[65385]: WARNING openstack [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.716504] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.717293] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.734028] env[65385]: INFO nova.compute.manager [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Took 43.58 seconds to build instance. [ 850.743266] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 850.758752] env[65385]: WARNING neutronclient.v2_0.client [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.759950] env[65385]: WARNING openstack [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.760086] env[65385]: WARNING openstack [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.929684] env[65385]: DEBUG nova.compute.manager [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 850.929788] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.931490] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b7050b-b955-443a-9d44-9570b9fa9eef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.941659] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.941939] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89ead30a-0bf6-419b-a089-51bfa7c0cd2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.955030] env[65385]: DEBUG oslo_vmware.api [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 850.955030] env[65385]: value = "task-4453802" [ 850.955030] env[65385]: _type = "Task" [ 850.955030] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.963108] env[65385]: DEBUG oslo_vmware.api [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.989175] env[65385]: DEBUG oslo_vmware.api [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453801, 'name': ReconfigVM_Task, 'duration_secs': 1.288099} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.989558] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Reconfigured VM instance instance-00000019 to attach disk [datastore2] 89f3c815-8671-47ce-9e74-bf6e652bb3c3/89f3c815-8671-47ce-9e74-bf6e652bb3c3.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.989987] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance '89f3c815-8671-47ce-9e74-bf6e652bb3c3' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 851.071566] env[65385]: DEBUG nova.network.neutron [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 851.105490] env[65385]: DEBUG nova.compute.utils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 851.119028] env[65385]: DEBUG nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 851.119028] env[65385]: DEBUG nova.network.neutron [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 851.119509] env[65385]: WARNING neutronclient.v2_0.client [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.120787] env[65385]: WARNING neutronclient.v2_0.client [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.120787] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.120787] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.134436] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d283b7ba-1559-4362-8efe-79944d465ad1 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.537s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 851.136058] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.290s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 851.136873] env[65385]: DEBUG nova.objects.instance [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lazy-loading 'resources' on Instance uuid ad29afd9-3e52-4fac-977f-91774acad7d0 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.236459] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202c1f12-c49a-42e2-b51e-e9a81b191d76 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.092s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 851.365387] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.366275] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.437243] env[65385]: DEBUG nova.network.neutron [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updated VIF entry in instance network info cache for port 7d96802a-1199-4a18-9a3d-f7b69a52590c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 851.437752] env[65385]: DEBUG nova.network.neutron [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updating instance_info_cache with network_info: [{"id": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "address": "fa:16:3e:42:06:97", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d96802a-11", "ovs_interfaceid": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 851.463944] env[65385]: DEBUG oslo_vmware.api [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453802, 'name': PowerOffVM_Task, 'duration_secs': 0.320799} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.464612] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.464612] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 851.464800] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23941227-91a8-497f-9dba-7d87a51bd53f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.479077] env[65385]: DEBUG nova.policy [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a4a60ae753f456cb987845b0a6b1afd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de072ae743844f6d9949acfe0fa6328c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 851.497939] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7c3177-bf4a-4a46-ab69-8d14ef02838e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.528213] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f842f7bd-d024-4c94-b3e1-148b3bad374e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.537513] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.537865] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.538139] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Deleting the datastore file [datastore1] 90c36f67-0c25-4487-a5e4-89827290953b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.559953] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9349996-3e6b-4b62-9382-e6ada7536607 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.562819] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance '89f3c815-8671-47ce-9e74-bf6e652bb3c3' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 851.574880] env[65385]: DEBUG oslo_vmware.api [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for the task: (returnval){ [ 851.574880] env[65385]: value = "task-4453804" [ 851.574880] env[65385]: _type = "Task" [ 851.574880] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.590234] env[65385]: DEBUG oslo_vmware.api [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453804, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.611944] env[65385]: DEBUG nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 851.942808] env[65385]: DEBUG oslo_concurrency.lockutils [req-435ab893-f66f-49c7-a439-27d131f9c973 req-7fd8c5c8-e8b5-420e-8cf1-df4e628ddd68 service nova] Releasing lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 852.007597] env[65385]: DEBUG nova.network.neutron [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Successfully updated port: d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 852.070809] env[65385]: WARNING neutronclient.v2_0.client [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.071185] env[65385]: WARNING neutronclient.v2_0.client [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.088810] env[65385]: DEBUG oslo_vmware.api [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Task: {'id': task-4453804, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153397} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.089108] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.089301] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 852.089481] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 852.089638] env[65385]: INFO nova.compute.manager [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Took 1.16 seconds to destroy the instance on the hypervisor. [ 852.089881] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 852.090097] env[65385]: DEBUG nova.compute.manager [-] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 852.090198] env[65385]: DEBUG nova.network.neutron [-] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 852.090439] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.090994] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 852.091285] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 852.129027] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb460240-5693-43a2-a3c7-0da4dc35bad4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.139021] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4474d1e1-0ebc-4bec-88d0-7717c1f9988d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.173115] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35fd1d7-14bd-41b3-a708-c765e8940f1a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.182075] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846ec6fe-5e2e-4b87-b679-57be83d17569 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.199346] env[65385]: DEBUG nova.compute.provider_tree [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.278666] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52913248-27d5-9d74-16ad-44e60466743c/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 852.280651] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36da8308-2574-4f8e-ba89-e27be89ca2a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.289413] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52913248-27d5-9d74-16ad-44e60466743c/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 852.289957] env[65385]: ERROR oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52913248-27d5-9d74-16ad-44e60466743c/disk-0.vmdk due to incomplete transfer. [ 852.290250] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b3025da4-2f5b-4d07-9840-c4b7ac4dde74 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.301748] env[65385]: DEBUG oslo_vmware.rw_handles [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52913248-27d5-9d74-16ad-44e60466743c/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 852.301915] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Uploaded image 68865751-553d-4087-a64b-e593c2630167 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 852.304470] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 852.304816] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7dd1aa94-620f-46c8-be39-3a8874011daa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.315747] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 852.315747] env[65385]: value = "task-4453805" [ 852.315747] env[65385]: _type = "Task" [ 852.315747] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.329611] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453805, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.344581] env[65385]: WARNING neutronclient.v2_0.client [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.345317] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 852.345658] env[65385]: WARNING openstack [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 852.377137] env[65385]: DEBUG nova.network.neutron [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Port 18c18d61-0493-40e3-8883-c90faabf147b binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 852.461304] env[65385]: DEBUG nova.network.neutron [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Successfully created port: f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 852.513155] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "refresh_cache-0c6206ee-c787-4cd1-9289-1b2620dc4c5f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.513155] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "refresh_cache-0c6206ee-c787-4cd1-9289-1b2620dc4c5f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.513155] env[65385]: DEBUG nova.network.neutron [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 852.524253] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.623037] env[65385]: DEBUG nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 852.669275] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 852.669275] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 852.669494] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 852.669698] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 852.669898] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 852.670077] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 852.670408] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.670633] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 852.670890] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 852.671226] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 852.671496] env[65385]: DEBUG nova.virt.hardware [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 852.673478] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712ead0f-28b0-4980-9d39-f35f74aac46a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.688510] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1cdead-94fc-44c8-b791-1059148f9f59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.702908] env[65385]: DEBUG nova.scheduler.client.report [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 852.826478] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453805, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.018177] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.018838] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 853.228783] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.093s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.231349] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.366s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.232926] env[65385]: INFO nova.compute.claims [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.257852] env[65385]: INFO nova.scheduler.client.report [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Deleted allocations for instance ad29afd9-3e52-4fac-977f-91774acad7d0 [ 853.327679] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453805, 'name': Destroy_Task, 'duration_secs': 0.617886} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.327679] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Destroyed the VM [ 853.327855] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 853.328352] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-449959db-0555-4012-aa66-30639359410c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.335927] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 853.335927] env[65385]: value = "task-4453806" [ 853.335927] env[65385]: _type = "Task" [ 853.335927] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.345075] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453806, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.409753] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.410260] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.410260] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.704265] env[65385]: DEBUG nova.network.neutron [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Updating instance_info_cache with network_info: [{"id": "97b8305b-a557-46c0-9f9a-f45d7ccd6e3a", "address": "fa:16:3e:26:4a:19", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97b8305b-a5", "ovs_interfaceid": "97b8305b-a557-46c0-9f9a-f45d7ccd6e3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 853.742167] env[65385]: DEBUG nova.network.neutron [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 853.773331] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a5d40403-6e71-43b0-8b38-5b9510b57ce3 tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.458s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.774503] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 21.669s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.777104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.777104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.777104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.779876] env[65385]: INFO nova.compute.manager [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Terminating instance [ 853.821606] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.822035] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 853.853401] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453806, 'name': RemoveSnapshot_Task, 'duration_secs': 0.405246} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.853401] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 853.853401] env[65385]: DEBUG nova.compute.manager [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 853.853401] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61fbb86-f3e9-44ab-b051-1caf91db20fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.901398] env[65385]: DEBUG nova.compute.manager [req-72e8db81-5144-499c-8d1d-181e9029b49a req-a0b6faf5-222b-4143-923b-4190b4dbc886 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Received event network-vif-plugged-97b8305b-a557-46c0-9f9a-f45d7ccd6e3a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 853.901498] env[65385]: DEBUG oslo_concurrency.lockutils [req-72e8db81-5144-499c-8d1d-181e9029b49a req-a0b6faf5-222b-4143-923b-4190b4dbc886 service nova] Acquiring lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.901688] env[65385]: DEBUG oslo_concurrency.lockutils [req-72e8db81-5144-499c-8d1d-181e9029b49a req-a0b6faf5-222b-4143-923b-4190b4dbc886 service nova] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.902266] env[65385]: DEBUG oslo_concurrency.lockutils [req-72e8db81-5144-499c-8d1d-181e9029b49a req-a0b6faf5-222b-4143-923b-4190b4dbc886 service nova] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.902266] env[65385]: DEBUG nova.compute.manager [req-72e8db81-5144-499c-8d1d-181e9029b49a req-a0b6faf5-222b-4143-923b-4190b4dbc886 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] No waiting events found dispatching network-vif-plugged-97b8305b-a557-46c0-9f9a-f45d7ccd6e3a {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 853.902517] env[65385]: WARNING nova.compute.manager [req-72e8db81-5144-499c-8d1d-181e9029b49a req-a0b6faf5-222b-4143-923b-4190b4dbc886 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Received unexpected event network-vif-plugged-97b8305b-a557-46c0-9f9a-f45d7ccd6e3a for instance with vm_state building and task_state spawning. [ 853.956391] env[65385]: WARNING neutronclient.v2_0.client [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 853.957132] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.957469] env[65385]: WARNING openstack [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.081753] env[65385]: DEBUG nova.network.neutron [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Updating instance_info_cache with network_info: [{"id": "d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e", "address": "fa:16:3e:48:6d:b4", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9c28a6e-bf", "ovs_interfaceid": "d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 854.110606] env[65385]: DEBUG nova.network.neutron [-] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 854.207320] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "refresh_cache-703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.207764] env[65385]: DEBUG nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Instance network_info: |[{"id": "97b8305b-a557-46c0-9f9a-f45d7ccd6e3a", "address": "fa:16:3e:26:4a:19", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97b8305b-a5", "ovs_interfaceid": "97b8305b-a557-46c0-9f9a-f45d7ccd6e3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 854.208272] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:4a:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4734e5e-2a76-4bda-8905-70c9bf9e007f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97b8305b-a557-46c0-9f9a-f45d7ccd6e3a', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.217038] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 854.217038] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.221339] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbfd2113-d670-4ba4-ac14-d485e33af3b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.257953] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.257953] env[65385]: value = "task-4453807" [ 854.257953] env[65385]: _type = "Task" [ 854.257953] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.268064] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453807, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.284387] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.285477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquired lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.285477] env[65385]: DEBUG nova.network.neutron [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 854.294821] env[65385]: DEBUG nova.network.neutron [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Successfully updated port: f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 854.366762] env[65385]: INFO nova.compute.manager [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Shelve offloading [ 854.421377] env[65385]: WARNING neutronclient.v2_0.client [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 854.495788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "6dae15ad-a340-47e3-ab6b-95b25043ca27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 854.495788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 854.523188] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.523188] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.523188] env[65385]: DEBUG nova.network.neutron [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 854.585386] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "refresh_cache-0c6206ee-c787-4cd1-9289-1b2620dc4c5f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.585757] env[65385]: DEBUG nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Instance network_info: |[{"id": "d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e", "address": "fa:16:3e:48:6d:b4", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9c28a6e-bf", "ovs_interfaceid": "d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 854.586269] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:6d:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9c28a6e-bfd5-46d3-96bc-ce3090c2a33e', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.597152] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 854.597415] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.598061] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-541ee544-0c22-44f0-96f0-28c512af05f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.617176] env[65385]: INFO nova.compute.manager [-] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Took 2.53 seconds to deallocate network for instance. [ 854.628862] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.628862] env[65385]: value = "task-4453808" [ 854.628862] env[65385]: _type = "Task" [ 854.628862] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.639266] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453808, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.750163] env[65385]: DEBUG nova.compute.manager [req-2a01060a-3196-4582-b079-0b3c0b725063 req-aaf905cb-40a9-4d4d-8107-4a557feaac65 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Received event network-vif-plugged-d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 854.750288] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a01060a-3196-4582-b079-0b3c0b725063 req-aaf905cb-40a9-4d4d-8107-4a557feaac65 service nova] Acquiring lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 854.750452] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a01060a-3196-4582-b079-0b3c0b725063 req-aaf905cb-40a9-4d4d-8107-4a557feaac65 service nova] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 854.750619] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a01060a-3196-4582-b079-0b3c0b725063 req-aaf905cb-40a9-4d4d-8107-4a557feaac65 service nova] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.750786] env[65385]: DEBUG nova.compute.manager [req-2a01060a-3196-4582-b079-0b3c0b725063 req-aaf905cb-40a9-4d4d-8107-4a557feaac65 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] No waiting events found dispatching network-vif-plugged-d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 854.750954] env[65385]: WARNING nova.compute.manager [req-2a01060a-3196-4582-b079-0b3c0b725063 req-aaf905cb-40a9-4d4d-8107-4a557feaac65 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Received unexpected event network-vif-plugged-d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e for instance with vm_state building and task_state spawning. [ 854.774234] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453807, 'name': CreateVM_Task, 'duration_secs': 0.419402} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.774234] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.774234] env[65385]: WARNING neutronclient.v2_0.client [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 854.774234] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.774234] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.774549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 854.774622] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f19ced2-a403-4f9a-b296-793e81e585cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.780981] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 854.780981] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528fbcdc-3469-524b-f463-6dea509aeadf" [ 854.780981] env[65385]: _type = "Task" [ 854.780981] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.796897] env[65385]: DEBUG nova.compute.utils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Can not refresh info_cache because instance was not found {{(pid=65385) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 854.797292] env[65385]: WARNING neutronclient.v2_0.client [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 854.797897] env[65385]: WARNING openstack [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.798625] env[65385]: WARNING openstack [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.807616] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528fbcdc-3469-524b-f463-6dea509aeadf, 'name': SearchDatastore_Task, 'duration_secs': 0.011846} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.811287] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "refresh_cache-9908d395-545d-4caf-9757-320d6253d61e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.811492] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquired lock "refresh_cache-9908d395-545d-4caf-9757-320d6253d61e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.811666] env[65385]: DEBUG nova.network.neutron [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 854.812845] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.816047] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.816047] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.816047] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.816047] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.816047] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee627deb-2aca-4e57-bf8c-ba8688bd8f58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.826173] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.826360] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.830276] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f5aacf0-05b8-4054-8daa-236fff16c576 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.833584] env[65385]: DEBUG nova.network.neutron [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 854.842023] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 854.842023] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529e3e6e-07fa-0c59-8b4f-dab5651b36d6" [ 854.842023] env[65385]: _type = "Task" [ 854.842023] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.845866] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3199c948-51d4-4fc1-9eb8-9b65e5fa653c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.857409] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529e3e6e-07fa-0c59-8b4f-dab5651b36d6, 'name': SearchDatastore_Task, 'duration_secs': 0.012411} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.860060] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6adc67e-9283-45cd-9a00-56408e7f4839 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.863439] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e8fe1c-3072-4037-92a8-e0f08b66228d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.870802] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 854.902396] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6bc9f767-ca4e-409c-81a5-4bb4781689bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.905182] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 854.905182] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52df99c2-94ef-6b58-8144-ba1f564994c8" [ 854.905182] env[65385]: _type = "Task" [ 854.905182] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.909547] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a7ba45-66c1-4cfb-9887-f8d7c566f794 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.920493] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 854.920493] env[65385]: value = "task-4453809" [ 854.920493] env[65385]: _type = "Task" [ 854.920493] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.929963] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52df99c2-94ef-6b58-8144-ba1f564994c8, 'name': SearchDatastore_Task, 'duration_secs': 0.010876} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.932377] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.932705] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f/703ed9b6-7cd8-4a84-9847-d34fb1c51a3f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.933280] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1435f0c2-d9fe-4645-9c88-214ca971e9ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.936257] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b75942-4167-4c0b-8474-8d24894fc33d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.944876] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 854.946386] env[65385]: DEBUG nova.compute.manager [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 854.946386] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af01e25-1dca-421b-99d3-9e694bede318 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.957669] env[65385]: DEBUG nova.compute.provider_tree [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.965432] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 854.965432] env[65385]: value = "task-4453810" [ 854.965432] env[65385]: _type = "Task" [ 854.965432] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.973087] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.973087] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.973200] env[65385]: DEBUG nova.network.neutron [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 854.981956] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453810, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.994637] env[65385]: DEBUG nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 855.032997] env[65385]: WARNING neutronclient.v2_0.client [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.033723] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.034150] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.129090] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.143685] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453808, 'name': CreateVM_Task, 'duration_secs': 0.365553} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.143936] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.144399] env[65385]: WARNING neutronclient.v2_0.client [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.144813] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.144994] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.145348] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 855.145615] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dba4b0dc-2295-4773-a489-a545731a76ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.154700] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 855.154700] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5266c282-8eb2-51b5-e3e2-e2e456b2f866" [ 855.154700] env[65385]: _type = "Task" [ 855.154700] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.165388] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5266c282-8eb2-51b5-e3e2-e2e456b2f866, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.316372] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.316657] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.468594] env[65385]: DEBUG nova.scheduler.client.report [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.476523] env[65385]: WARNING neutronclient.v2_0.client [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.477233] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.477544] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.497788] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453810, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503006} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.501274] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f/703ed9b6-7cd8-4a84-9847-d34fb1c51a3f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.501274] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.503943] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20786656-0c2e-47a6-8118-6f2ae3163104 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.513145] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 855.513145] env[65385]: value = "task-4453811" [ 855.513145] env[65385]: _type = "Task" [ 855.513145] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.521273] env[65385]: DEBUG nova.network.neutron [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 855.525897] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.530312] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.631853] env[65385]: DEBUG nova.network.neutron [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 855.668358] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5266c282-8eb2-51b5-e3e2-e2e456b2f866, 'name': SearchDatastore_Task, 'duration_secs': 0.067631} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.668822] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.668822] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.668965] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.669120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.669289] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.669756] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-502ef6c4-da2d-4911-bc63-61ead65f0965 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.680713] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.680910] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.681678] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-905535a5-47e1-4b44-9663-63af4277fab5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.688787] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 855.688787] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5233873a-9dd3-bd6a-168c-30bdd6b47184" [ 855.688787] env[65385]: _type = "Task" [ 855.688787] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.700107] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.700800] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.707241] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5233873a-9dd3-bd6a-168c-30bdd6b47184, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.918018] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.918018] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.935614] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.936172] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.977912] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.746s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 855.977912] env[65385]: DEBUG nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 855.983052] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.649s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.983052] env[65385]: DEBUG nova.objects.instance [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lazy-loading 'resources' on Instance uuid 0673a0d2-76ac-4cd5-8cc3-b8596877c641 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.006569] env[65385]: WARNING neutronclient.v2_0.client [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.007276] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.007630] env[65385]: WARNING openstack [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.027829] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Releasing lock "refresh_cache-ad29afd9-3e52-4fac-977f-91774acad7d0" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.028292] env[65385]: DEBUG nova.compute.manager [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 856.028483] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 856.029105] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072597} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.029105] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8fa0a44-55cc-46d1-af8d-3b82b82affec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.031246] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.031933] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9234bdf8-c6e9-43c9-b6c8-8eac9a2f2ac6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.060372] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f/703ed9b6-7cd8-4a84-9847-d34fb1c51a3f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.066722] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1283d9b2-1288-4400-b030-329e7627926e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.092318] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ccb59b-7b86-4463-aeb2-7a56fda5a774 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.111538] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 856.111538] env[65385]: value = "task-4453812" [ 856.111538] env[65385]: _type = "Task" [ 856.111538] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.135624] env[65385]: WARNING nova.virt.vmwareapi.vmops [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad29afd9-3e52-4fac-977f-91774acad7d0 could not be found. [ 856.136901] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.136901] env[65385]: INFO nova.compute.manager [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Took 0.11 seconds to destroy the instance on the hypervisor. [ 856.137900] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 856.142601] env[65385]: DEBUG nova.compute.manager [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 856.142742] env[65385]: DEBUG nova.network.neutron [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 856.142981] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.143556] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.143952] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.151585] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453812, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.167922] env[65385]: WARNING neutronclient.v2_0.client [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.168641] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.169075] env[65385]: WARNING openstack [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.201311] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5233873a-9dd3-bd6a-168c-30bdd6b47184, 'name': SearchDatastore_Task, 'duration_secs': 0.009719} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.202394] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bedfecb-3a01-4364-a8e5-22e889882425 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.211050] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 856.211050] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525a059b-771b-ea00-5363-58694848a79c" [ 856.211050] env[65385]: _type = "Task" [ 856.211050] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.215240] env[65385]: WARNING neutronclient.v2_0.client [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.216072] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.216352] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.232840] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525a059b-771b-ea00-5363-58694848a79c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.279262] env[65385]: DEBUG nova.network.neutron [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 856.279624] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.320531] env[65385]: DEBUG nova.network.neutron [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Updating instance_info_cache with network_info: [{"id": "f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55", "address": "fa:16:3e:d5:05:ad", "network": {"id": "84b6a9d4-a34c-4839-8959-accb4a12f85b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1972478708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de072ae743844f6d9949acfe0fa6328c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6b500f6-8c", "ovs_interfaceid": "f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.446752] env[65385]: DEBUG nova.network.neutron [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.488214] env[65385]: DEBUG nova.compute.utils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 856.494378] env[65385]: DEBUG nova.network.neutron [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Updating instance_info_cache with network_info: [{"id": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "address": "fa:16:3e:f3:7a:17", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd15afa85-05", "ovs_interfaceid": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.496571] env[65385]: DEBUG nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 856.496970] env[65385]: DEBUG nova.network.neutron [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 856.497419] env[65385]: WARNING neutronclient.v2_0.client [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.497912] env[65385]: WARNING neutronclient.v2_0.client [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.498630] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.499305] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.628069] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453812, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.725981] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525a059b-771b-ea00-5363-58694848a79c, 'name': SearchDatastore_Task, 'duration_secs': 0.011816} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.726470] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.726729] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0c6206ee-c787-4cd1-9289-1b2620dc4c5f/0c6206ee-c787-4cd1-9289-1b2620dc4c5f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.727072] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3f90c79-308f-405b-b102-350c1b7dc3db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.740166] env[65385]: DEBUG nova.policy [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2a287457352470794887b229ea90cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60de0d1c162342209795a180391661f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 856.743517] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 856.743517] env[65385]: value = "task-4453813" [ 856.743517] env[65385]: _type = "Task" [ 856.743517] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.756475] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.784228] env[65385]: DEBUG nova.network.neutron [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.823704] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Releasing lock "refresh_cache-9908d395-545d-4caf-9757-320d6253d61e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.824219] env[65385]: DEBUG nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Instance network_info: |[{"id": "f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55", "address": "fa:16:3e:d5:05:ad", "network": {"id": "84b6a9d4-a34c-4839-8959-accb4a12f85b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1972478708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de072ae743844f6d9949acfe0fa6328c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6b500f6-8c", "ovs_interfaceid": "f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 856.825724] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:05:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604056d6-6dd6-47fa-9eaa-6863a3a7c488', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.835477] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Creating folder: Project (de072ae743844f6d9949acfe0fa6328c). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.841244] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40944852-77f8-4f54-88a9-ca98f4e1d5b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.856960] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Created folder: Project (de072ae743844f6d9949acfe0fa6328c) in parent group-v870881. [ 856.856960] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Creating folder: Instances. Parent ref: group-v871066. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.856960] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a58819f-43c6-40ab-8d86-11a8812e4e41 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.875506] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Created folder: Instances in parent group-v871066. [ 856.875789] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 856.876048] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.876285] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a07ec21-deca-4a30-a7a9-c50bf5155494 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.904098] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "26a2d464-0aab-463f-8265-948a4f0ff188" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.904098] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.907154] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.907154] env[65385]: value = "task-4453816" [ 856.907154] env[65385]: _type = "Task" [ 856.907154] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.921211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "fa9a5278-1477-485a-9201-a37187488aef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.921211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "fa9a5278-1477-485a-9201-a37187488aef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.929794] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453816, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.949887] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.995330] env[65385]: DEBUG nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 856.998207] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.999220] env[65385]: WARNING neutronclient.v2_0.client [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.999430] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.999884] env[65385]: WARNING openstack [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 857.006030] env[65385]: WARNING neutronclient.v2_0.client [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 857.010583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f11d5cf-9ff5-49b1-a18c-2e7ef5a1c875 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.024063] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f433f27e-4b69-4e9b-a931-7553da26dc86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.063625] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5bde27-f0d4-4d3c-b1d1-e6f5fb78d01b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.078228] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a308bf66-a06f-4632-8eb3-b41b11e79bf7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.098912] env[65385]: DEBUG nova.compute.provider_tree [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.137570] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453812, 'name': ReconfigVM_Task, 'duration_secs': 0.726084} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.137570] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f/703ed9b6-7cd8-4a84-9847-d34fb1c51a3f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.137570] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3fb764cc-32b7-4a96-954a-051439c21021 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.147945] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 857.147945] env[65385]: value = "task-4453817" [ 857.147945] env[65385]: _type = "Task" [ 857.147945] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.160751] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453817, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.202469] env[65385]: DEBUG nova.network.neutron [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Successfully created port: 722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 857.256107] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453813, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.287787] env[65385]: INFO nova.compute.manager [-] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Took 1.14 seconds to deallocate network for instance. [ 857.409641] env[65385]: DEBUG nova.compute.utils [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 857.429067] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453816, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.445126] env[65385]: DEBUG nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 857.495185] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a539f8e-30e6-4378-95df-43ed03a3ca37 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.524187] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d5b91a-5cc5-4195-8b8d-f40c6e693650 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.527558] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 857.528843] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba911536-cd31-4f2d-8b57-d2d7fdf77d7e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.538730] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance '89f3c815-8671-47ce-9e74-bf6e652bb3c3' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 857.546038] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 857.546038] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73f59c6e-ba06-45e3-96c6-11dc9c196f4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.604959] env[65385]: DEBUG nova.scheduler.client.report [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.630548] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 857.631172] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 857.631172] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleting the datastore file [datastore2] b1cca536-e1d9-43f8-8a5c-be3981c106e7 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.631335] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59f6c4be-9bdc-4350-8d1d-cc04631b0553 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.641459] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 857.641459] env[65385]: value = "task-4453819" [ 857.641459] env[65385]: _type = "Task" [ 857.641459] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.652325] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.662450] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453817, 'name': Rename_Task, 'duration_secs': 0.258043} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.662910] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.663114] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8132ec89-f97a-4eba-ac77-0c81a78085ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.671313] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 857.671313] env[65385]: value = "task-4453820" [ 857.671313] env[65385]: _type = "Task" [ 857.671313] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.682468] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453820, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.754523] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535852} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.754777] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 0c6206ee-c787-4cd1-9289-1b2620dc4c5f/0c6206ee-c787-4cd1-9289-1b2620dc4c5f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.754965] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.755233] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae6b649b-4439-40e8-9e76-657dbfedefb3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.762526] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 857.762526] env[65385]: value = "task-4453821" [ 857.762526] env[65385]: _type = "Task" [ 857.762526] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.771491] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453821, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.795274] env[65385]: INFO nova.compute.manager [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance disappeared during terminate [ 857.795513] env[65385]: DEBUG oslo_concurrency.lockutils [None req-add5e063-61af-473c-919d-b147465db72c tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "ad29afd9-3e52-4fac-977f-91774acad7d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.021s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.919163] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.015s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.923254] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453816, 'name': CreateVM_Task, 'duration_secs': 0.63838} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.924100] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.924684] env[65385]: WARNING neutronclient.v2_0.client [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 857.925063] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.925218] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 857.925524] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 857.926070] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d20e931-c4d0-4e25-bd46-1f397b425278 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.931662] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 857.931662] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521de244-d432-f7a2-52b1-5c6305b30c49" [ 857.931662] env[65385]: _type = "Task" [ 857.931662] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.946299] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521de244-d432-f7a2-52b1-5c6305b30c49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.962802] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 858.032673] env[65385]: DEBUG nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 858.048377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a8db0f-c430-4416-898f-2aa50805d5fd tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance '89f3c815-8671-47ce-9e74-bf6e652bb3c3' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 858.062289] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 858.062558] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 858.062732] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 858.062881] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 858.063022] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 858.063218] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 858.063420] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.063588] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 858.063768] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 858.063913] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 858.064152] env[65385]: DEBUG nova.virt.hardware [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 858.065059] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bad965-e7da-4aa5-8dff-57ba2d3baded {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.074710] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7ce3ed-f7b3-4b98-b208-913749dd4726 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.111127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.129s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 858.115442] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.420s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 858.117191] env[65385]: INFO nova.compute.claims [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.147367] env[65385]: INFO nova.scheduler.client.report [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted allocations for instance 0673a0d2-76ac-4cd5-8cc3-b8596877c641 [ 858.154945] env[65385]: DEBUG oslo_vmware.api [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138593} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.155620] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.155813] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 858.155983] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 858.186551] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453820, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.187763] env[65385]: INFO nova.scheduler.client.report [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted allocations for instance b1cca536-e1d9-43f8-8a5c-be3981c106e7 [ 858.273960] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453821, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07205} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.273960] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.274504] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1e359e-0907-4169-a5ac-4ddd049b5d1a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.299826] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 0c6206ee-c787-4cd1-9289-1b2620dc4c5f/0c6206ee-c787-4cd1-9289-1b2620dc4c5f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.300484] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8eacd387-1f2b-4b4a-887f-1d9ab894ac69 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.325513] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 858.325513] env[65385]: value = "task-4453822" [ 858.325513] env[65385]: _type = "Task" [ 858.325513] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.335508] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453822, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.444995] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521de244-d432-f7a2-52b1-5c6305b30c49, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.445363] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 858.445654] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.446028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.446131] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.446320] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.446635] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e437b7d5-543e-4b4e-a576-e4071e25b4f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.458751] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.458751] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.459555] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91e35fb0-b7a1-48ce-b96e-5db683aaec2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.466616] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 858.466616] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ff33aa-f1cc-d6cc-7163-df3ad307fcd0" [ 858.466616] env[65385]: _type = "Task" [ 858.466616] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.477106] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ff33aa-f1cc-d6cc-7163-df3ad307fcd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.655686] env[65385]: DEBUG oslo_concurrency.lockutils [None req-19159113-6b71-4736-8f4a-ad0c7fc3b0ab tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "0673a0d2-76ac-4cd5-8cc3-b8596877c641" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.883s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 858.683707] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453820, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.692877] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 858.839772] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453822, 'name': ReconfigVM_Task, 'duration_secs': 0.334874} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.840068] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 0c6206ee-c787-4cd1-9289-1b2620dc4c5f/0c6206ee-c787-4cd1-9289-1b2620dc4c5f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.840767] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0a504bd-b241-4c76-a9d9-5b3b905e2236 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.849240] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 858.849240] env[65385]: value = "task-4453823" [ 858.849240] env[65385]: _type = "Task" [ 858.849240] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.860671] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453823, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.940902] env[65385]: DEBUG nova.network.neutron [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Successfully updated port: 722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 858.980397] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ff33aa-f1cc-d6cc-7163-df3ad307fcd0, 'name': SearchDatastore_Task, 'duration_secs': 0.011418} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.981442] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c2908fa-8e93-474e-b935-cd759d41ea01 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.989680] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 858.989680] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523dcd6f-e3b4-b39f-1c7e-a6ea5e5793c0" [ 858.989680] env[65385]: _type = "Task" [ 858.989680] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.001521] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523dcd6f-e3b4-b39f-1c7e-a6ea5e5793c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.043122] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "26a2d464-0aab-463f-8265-948a4f0ff188" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 859.043837] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.044422] env[65385]: INFO nova.compute.manager [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Attaching volume 33fd2b93-a90e-4837-9089-e9d5cab1e2df to /dev/sdb [ 859.105442] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f092e92-8e0e-43fa-9ae5-3675fe8e01be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.113855] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ad289e-d4db-460a-b877-0650a35e0b12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.137922] env[65385]: DEBUG nova.virt.block_device [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Updating existing volume attachment record: 4e0da0d8-6c7c-4362-a283-5bc4e1c8e2d0 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 859.186857] env[65385]: DEBUG oslo_vmware.api [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453820, 'name': PowerOnVM_Task, 'duration_secs': 1.210818} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.187178] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.187394] env[65385]: INFO nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Took 12.00 seconds to spawn the instance on the hypervisor. [ 859.187569] env[65385]: DEBUG nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 859.188412] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff73d6a-a60a-4b09-8f7c-61cf55b4e54b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.261629] env[65385]: DEBUG nova.compute.manager [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Received event network-changed-97b8305b-a557-46c0-9f9a-f45d7ccd6e3a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 859.262409] env[65385]: DEBUG nova.compute.manager [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Refreshing instance network info cache due to event network-changed-97b8305b-a557-46c0-9f9a-f45d7ccd6e3a. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 859.265324] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Acquiring lock "refresh_cache-703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.265477] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Acquired lock "refresh_cache-703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 859.265632] env[65385]: DEBUG nova.network.neutron [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Refreshing network info cache for port 97b8305b-a557-46c0-9f9a-f45d7ccd6e3a {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 859.371226] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453823, 'name': Rename_Task, 'duration_secs': 0.364791} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.371369] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 859.371625] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a43db0a1-81ab-4b7c-8e14-950067941c3b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.379360] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 859.379360] env[65385]: value = "task-4453826" [ 859.379360] env[65385]: _type = "Task" [ 859.379360] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.402479] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453826, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.402479] env[65385]: DEBUG nova.compute.manager [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Received event network-changed-d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 859.402479] env[65385]: DEBUG nova.compute.manager [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Refreshing instance network info cache due to event network-changed-d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 859.402479] env[65385]: DEBUG oslo_concurrency.lockutils [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Acquiring lock "refresh_cache-0c6206ee-c787-4cd1-9289-1b2620dc4c5f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.402479] env[65385]: DEBUG oslo_concurrency.lockutils [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Acquired lock "refresh_cache-0c6206ee-c787-4cd1-9289-1b2620dc4c5f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 859.402479] env[65385]: DEBUG nova.network.neutron [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Refreshing network info cache for port d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 859.444082] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.444312] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 859.444485] env[65385]: DEBUG nova.network.neutron [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 859.503063] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523dcd6f-e3b4-b39f-1c7e-a6ea5e5793c0, 'name': SearchDatastore_Task, 'duration_secs': 0.013278} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.503409] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.503683] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 9908d395-545d-4caf-9757-320d6253d61e/9908d395-545d-4caf-9757-320d6253d61e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.503966] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ccb0bf70-87f1-45fd-bb11-fdf2783b0fc8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.526378] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 859.526378] env[65385]: value = "task-4453828" [ 859.526378] env[65385]: _type = "Task" [ 859.526378] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.536561] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453828, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.717759] env[65385]: INFO nova.compute.manager [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Took 34.52 seconds to build instance. [ 859.747729] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c801aa23-d008-48db-a4df-e0d7f6977a9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.757893] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31aed677-ce37-4ff9-8289-ee17681627b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.795114] env[65385]: WARNING neutronclient.v2_0.client [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 859.795803] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.796172] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.804548] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 859.804823] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.805012] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 859.805235] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.805379] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.808233] env[65385]: INFO nova.compute.manager [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Terminating instance [ 859.810773] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f817bc7-85ef-46f4-abce-f88b8a881d9e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.822373] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cf8e8c-01ba-453b-9020-268ccffc204e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.842931] env[65385]: DEBUG nova.compute.provider_tree [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.892773] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453826, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.907632] env[65385]: WARNING neutronclient.v2_0.client [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 859.908363] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.908748] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.948439] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.948787] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.022619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "7fcbdc79-688c-479f-94e2-f4542abe714b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.022874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.023096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "7fcbdc79-688c-479f-94e2-f4542abe714b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.023278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.023441] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.025783] env[65385]: INFO nova.compute.manager [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Terminating instance [ 860.039534] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453828, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.220639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3c687d6b-507c-4f3c-aca1-78014fdc28b1 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.040s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.229025] env[65385]: DEBUG nova.network.neutron [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 860.316879] env[65385]: DEBUG nova.compute.manager [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 860.317166] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.318081] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c132c5-ed6f-4f74-8240-673dd4875574 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.323034] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.323429] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.336756] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.337063] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83e00b72-bde4-4aba-bf9b-acac17e5d11a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.348050] env[65385]: DEBUG nova.scheduler.client.report [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 860.351330] env[65385]: DEBUG oslo_vmware.api [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 860.351330] env[65385]: value = "task-4453829" [ 860.351330] env[65385]: _type = "Task" [ 860.351330] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.357357] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.357727] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.370448] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.370862] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.383848] env[65385]: DEBUG oslo_vmware.api [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453829, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.394379] env[65385]: DEBUG oslo_vmware.api [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453826, 'name': PowerOnVM_Task, 'duration_secs': 0.555998} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.401784] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.402164] env[65385]: INFO nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Took 10.50 seconds to spawn the instance on the hypervisor. [ 860.402604] env[65385]: DEBUG nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 860.404427] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7ce05b-8345-4630-9018-06f9332d6c07 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.514276] env[65385]: WARNING neutronclient.v2_0.client [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.514933] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.515537] env[65385]: WARNING openstack [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.539338] env[65385]: DEBUG nova.compute.manager [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 860.539338] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.539338] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac385be-bf65-43f2-82da-7ec1d90dbb31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.546867] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453828, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519167} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.549770] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 9908d395-545d-4caf-9757-320d6253d61e/9908d395-545d-4caf-9757-320d6253d61e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.549770] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.549961] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.551261] env[65385]: WARNING neutronclient.v2_0.client [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.552045] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.552451] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.560904] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7c3f96c8-4e31-4c21-8049-915f4c674f82 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.562922] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-335662ac-0dcd-416f-aa44-bd5f20697b55 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.566660] env[65385]: WARNING neutronclient.v2_0.client [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.567435] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.567857] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.582670] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 860.582670] env[65385]: value = "task-4453831" [ 860.582670] env[65385]: _type = "Task" [ 860.582670] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.584240] env[65385]: DEBUG oslo_vmware.api [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 860.584240] env[65385]: value = "task-4453830" [ 860.584240] env[65385]: _type = "Task" [ 860.584240] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.599980] env[65385]: DEBUG oslo_vmware.api [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.600482] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453831, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.727055] env[65385]: DEBUG nova.network.neutron [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Updating instance_info_cache with network_info: [{"id": "722cc380-0268-4488-ad6c-462f3af7b938", "address": "fa:16:3e:a9:5f:40", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap722cc380-02", "ovs_interfaceid": "722cc380-0268-4488-ad6c-462f3af7b938", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 860.753839] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "75ba706f-2e72-4f84-b02f-db4381951e77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.753839] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "75ba706f-2e72-4f84-b02f-db4381951e77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.754070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "75ba706f-2e72-4f84-b02f-db4381951e77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.754196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "75ba706f-2e72-4f84-b02f-db4381951e77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.754684] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "75ba706f-2e72-4f84-b02f-db4381951e77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.756816] env[65385]: INFO nova.compute.manager [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Terminating instance [ 860.771464] env[65385]: DEBUG nova.network.neutron [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Updated VIF entry in instance network info cache for port 97b8305b-a557-46c0-9f9a-f45d7ccd6e3a. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 860.772047] env[65385]: DEBUG nova.network.neutron [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Updating instance_info_cache with network_info: [{"id": "97b8305b-a557-46c0-9f9a-f45d7ccd6e3a", "address": "fa:16:3e:26:4a:19", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97b8305b-a5", "ovs_interfaceid": "97b8305b-a557-46c0-9f9a-f45d7ccd6e3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 860.794032] env[65385]: DEBUG nova.network.neutron [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Updated VIF entry in instance network info cache for port d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 860.794443] env[65385]: DEBUG nova.network.neutron [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Updating instance_info_cache with network_info: [{"id": "d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e", "address": "fa:16:3e:48:6d:b4", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9c28a6e-bf", "ovs_interfaceid": "d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 860.859032] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.744s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.859605] env[65385]: DEBUG nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 860.862371] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.390s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.862611] env[65385]: DEBUG nova.objects.instance [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lazy-loading 'resources' on Instance uuid f25cc41e-560e-4cab-8bf0-dac47899d037 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.873325] env[65385]: DEBUG oslo_vmware.api [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453829, 'name': PowerOffVM_Task, 'duration_secs': 0.379579} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.874555] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.874881] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.875163] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97bd5a5e-921c-414a-828e-f978074b8cc1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.927118] env[65385]: INFO nova.compute.manager [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Took 35.28 seconds to build instance. [ 860.963771] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.963835] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.963986] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleting the datastore file [datastore1] 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.964321] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfad2fed-624a-4813-8fe0-e1af52da3259 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.973130] env[65385]: DEBUG oslo_vmware.api [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 860.973130] env[65385]: value = "task-4453833" [ 860.973130] env[65385]: _type = "Task" [ 860.973130] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.985890] env[65385]: DEBUG oslo_vmware.api [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.098619] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453831, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109883} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.102060] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.102456] env[65385]: DEBUG oslo_vmware.api [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453830, 'name': PowerOffVM_Task, 'duration_secs': 0.283294} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.103221] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f5fa55-99f6-4382-a103-effb6d80dfb3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.106135] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 861.106309] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 861.106588] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46134307-918c-46f5-9721-5b3f965440e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.129784] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 9908d395-545d-4caf-9757-320d6253d61e/9908d395-545d-4caf-9757-320d6253d61e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.130106] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80f2b28a-03ef-44d2-be6a-fd88613ebac4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.159295] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 861.159295] env[65385]: value = "task-4453835" [ 861.159295] env[65385]: _type = "Task" [ 861.159295] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.171043] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453835, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.181638] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 861.181880] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 861.182028] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Deleting the datastore file [datastore2] 7fcbdc79-688c-479f-94e2-f4542abe714b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.182402] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6f5e0ed-382e-45a4-a1ec-c2544adf4c05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.193068] env[65385]: DEBUG oslo_vmware.api [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for the task: (returnval){ [ 861.193068] env[65385]: value = "task-4453836" [ 861.193068] env[65385]: _type = "Task" [ 861.193068] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.203069] env[65385]: DEBUG oslo_vmware.api [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.230425] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.230979] env[65385]: DEBUG nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Instance network_info: |[{"id": "722cc380-0268-4488-ad6c-462f3af7b938", "address": "fa:16:3e:a9:5f:40", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap722cc380-02", "ovs_interfaceid": "722cc380-0268-4488-ad6c-462f3af7b938", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 861.231756] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:5f:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '722cc380-0268-4488-ad6c-462f3af7b938', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.245745] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 861.246701] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.247448] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2400496-591c-41c4-8d5c-cf3c3e72ee4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.272371] env[65385]: DEBUG nova.compute.manager [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 861.272842] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.273325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0668c1bd-cb6c-4b3a-a3e9-044d196ca29d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.277662] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Releasing lock "refresh_cache-703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.277842] env[65385]: DEBUG nova.compute.manager [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Received event network-vif-plugged-f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 861.277971] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Acquiring lock "9908d395-545d-4caf-9757-320d6253d61e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.278198] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Lock "9908d395-545d-4caf-9757-320d6253d61e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.278433] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Lock "9908d395-545d-4caf-9757-320d6253d61e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.278810] env[65385]: DEBUG nova.compute.manager [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] No waiting events found dispatching network-vif-plugged-f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 861.278810] env[65385]: WARNING nova.compute.manager [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Received unexpected event network-vif-plugged-f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 for instance with vm_state building and task_state spawning. [ 861.279118] env[65385]: DEBUG nova.compute.manager [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Received event network-changed-f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 861.279118] env[65385]: DEBUG nova.compute.manager [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Refreshing instance network info cache due to event network-changed-f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 861.279267] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Acquiring lock "refresh_cache-9908d395-545d-4caf-9757-320d6253d61e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.279443] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Acquired lock "refresh_cache-9908d395-545d-4caf-9757-320d6253d61e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.279664] env[65385]: DEBUG nova.network.neutron [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Refreshing network info cache for port f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 861.283920] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.283920] env[65385]: value = "task-4453837" [ 861.283920] env[65385]: _type = "Task" [ 861.283920] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.290577] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.291864] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c568db76-20d5-4a57-a790-10e95d6c2fd0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.300952] env[65385]: DEBUG oslo_concurrency.lockutils [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Releasing lock "refresh_cache-0c6206ee-c787-4cd1-9289-1b2620dc4c5f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.301825] env[65385]: DEBUG nova.compute.manager [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Received event network-vif-deleted-c2cde5c4-0094-40e7-965c-60977c968a4d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 861.301825] env[65385]: DEBUG nova.compute.manager [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received event network-changed-7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 861.301825] env[65385]: DEBUG nova.compute.manager [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Refreshing instance network info cache due to event network-changed-7d96802a-1199-4a18-9a3d-f7b69a52590c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 861.302233] env[65385]: DEBUG oslo_concurrency.lockutils [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Acquiring lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.302233] env[65385]: DEBUG oslo_concurrency.lockutils [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Acquired lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.302368] env[65385]: DEBUG nova.network.neutron [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Refreshing network info cache for port 7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 861.304276] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453837, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.306528] env[65385]: DEBUG oslo_vmware.api [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 861.306528] env[65385]: value = "task-4453838" [ 861.306528] env[65385]: _type = "Task" [ 861.306528] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.317283] env[65385]: DEBUG oslo_vmware.api [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453838, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.369929] env[65385]: DEBUG nova.compute.utils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 861.374015] env[65385]: DEBUG nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 861.374015] env[65385]: DEBUG nova.network.neutron [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 861.374015] env[65385]: WARNING neutronclient.v2_0.client [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.374015] env[65385]: WARNING neutronclient.v2_0.client [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.374015] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.374015] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.429301] env[65385]: DEBUG oslo_concurrency.lockutils [None req-decaa873-f873-4e7f-b98e-7cc083759345 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.794s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.444138] env[65385]: DEBUG nova.policy [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f10737b4a00442ab3abd1d126323ae8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc464c9d4e49499ba42b61ce293ac4fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 861.492276] env[65385]: DEBUG oslo_vmware.api [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160985} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.492276] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.492276] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.492276] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.492276] env[65385]: INFO nova.compute.manager [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Took 1.17 seconds to destroy the instance on the hypervisor. [ 861.492276] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 861.493302] env[65385]: DEBUG nova.compute.manager [-] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 861.493302] env[65385]: DEBUG nova.network.neutron [-] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 861.493696] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.494814] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.495075] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.580718] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.652910] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.675537] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453835, 'name': ReconfigVM_Task, 'duration_secs': 0.388831} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.675727] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 9908d395-545d-4caf-9757-320d6253d61e/9908d395-545d-4caf-9757-320d6253d61e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.676304] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a32e3cdb-f436-4441-9314-71425c325da6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.690136] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 861.690136] env[65385]: value = "task-4453840" [ 861.690136] env[65385]: _type = "Task" [ 861.690136] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.704347] env[65385]: DEBUG oslo_vmware.api [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Task: {'id': task-4453836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168491} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.708616] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.709641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.709985] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.710323] env[65385]: INFO nova.compute.manager [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 861.710692] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 861.711319] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453840, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.715433] env[65385]: DEBUG nova.compute.manager [-] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 861.715614] env[65385]: DEBUG nova.network.neutron [-] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 861.715881] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.716586] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.717114] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.783012] env[65385]: WARNING neutronclient.v2_0.client [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.783434] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.783558] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.803957] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453837, 'name': CreateVM_Task, 'duration_secs': 0.447156} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.804248] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 861.804806] env[65385]: WARNING neutronclient.v2_0.client [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.805443] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.805792] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.815942] env[65385]: WARNING neutronclient.v2_0.client [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.816494] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.816681] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.817227] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 861.817636] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72160b33-c0c0-4b3f-aa0e-8a3dce5e0687 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.834367] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 861.834367] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523d4ef1-1872-74d1-45f0-527946d96aa0" [ 861.834367] env[65385]: _type = "Task" [ 861.834367] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.835746] env[65385]: DEBUG oslo_vmware.api [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453838, 'name': PowerOffVM_Task, 'duration_secs': 0.310374} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.840646] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 861.841156] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 861.841156] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1723839-4fd6-454c-8e8b-3f7ff94941d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.853317] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523d4ef1-1872-74d1-45f0-527946d96aa0, 'name': SearchDatastore_Task, 'duration_secs': 0.013262} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.853884] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.854536] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.854641] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.854809] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.855112] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.855671] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84c8127a-9438-423b-8650-7ac9ac2abd94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.860032] env[65385]: DEBUG nova.network.neutron [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Successfully created port: 3f2342b2-bc3c-489f-a61b-ffb18a20cce6 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 861.877053] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.877053] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.877053] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a329285-c4e7-4da4-8160-715700ccbd9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.883684] env[65385]: DEBUG nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 861.886489] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 861.886489] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52734b44-5332-8715-210b-7882daf472a1" [ 861.886489] env[65385]: _type = "Task" [ 861.886489] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.897742] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52734b44-5332-8715-210b-7882daf472a1, 'name': SearchDatastore_Task, 'duration_secs': 0.010749} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.898372] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a83c48f-8c3d-4476-9844-45c658ea7089 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.908499] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 861.908499] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523634d6-8fbb-c4b4-2f57-2526a3656b54" [ 861.908499] env[65385]: _type = "Task" [ 861.908499] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.920863] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523634d6-8fbb-c4b4-2f57-2526a3656b54, 'name': SearchDatastore_Task, 'duration_secs': 0.011318} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.924026] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.924026] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.924026] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 861.924026] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 861.924026] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleting the datastore file [datastore1] 75ba706f-2e72-4f84-b02f-db4381951e77 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.924026] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a839e872-f1b4-41b9-9dbf-6c1ff993549d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.926109] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03b6a7be-df38-4069-8458-ceee0d8145bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.942454] env[65385]: DEBUG oslo_vmware.api [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 861.942454] env[65385]: value = "task-4453843" [ 861.942454] env[65385]: _type = "Task" [ 861.942454] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.942454] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 861.942454] env[65385]: value = "task-4453842" [ 861.942454] env[65385]: _type = "Task" [ 861.942454] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.958099] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.985395] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2619813-c06b-40d2-82cf-5d179b6776ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.996376] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03591ea-2790-48c2-ba46-02c21a05410b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.033219] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.035576] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9306fac-bec6-4a27-afed-2c7f4ad0bcc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.046833] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd54f434-62fe-4dbe-8a54-cc9b6dc75e8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.062857] env[65385]: DEBUG nova.compute.provider_tree [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.202322] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453840, 'name': Rename_Task, 'duration_secs': 0.172808} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.202679] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.202990] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08d5a0d8-5820-4323-b824-ce0359662252 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.211534] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 862.211534] env[65385]: value = "task-4453844" [ 862.211534] env[65385]: _type = "Task" [ 862.211534] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.222270] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453844, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.457569] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "cf37e10a-14f4-4456-bad0-d7528457658d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.457832] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.470282] env[65385]: DEBUG oslo_vmware.api [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213705} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.470282] env[65385]: DEBUG nova.network.neutron [-] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 862.470282] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453842, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522727} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.470282] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.470768] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 862.470768] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 862.470889] env[65385]: INFO nova.compute.manager [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Took 1.20 seconds to destroy the instance on the hypervisor. [ 862.471245] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 862.474196] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.474196] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.474196] env[65385]: DEBUG nova.compute.manager [-] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 862.474196] env[65385]: DEBUG nova.network.neutron [-] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 862.474196] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.474196] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.474762] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.483405] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3be6f7d-6496-459e-8bec-6e883c3146d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.501088] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 862.501088] env[65385]: value = "task-4453845" [ 862.501088] env[65385]: _type = "Task" [ 862.501088] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.512797] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453845, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.568514] env[65385]: DEBUG nova.scheduler.client.report [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 862.618332] env[65385]: DEBUG nova.compute.manager [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 862.619979] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.622622] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7d629a-60be-4dff-bb32-c04f64403724 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.637817] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.638228] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.678108] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.678514] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.723714] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453844, 'name': PowerOnVM_Task} progress is 98%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.876192] env[65385]: DEBUG nova.network.neutron [-] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 862.899504] env[65385]: DEBUG nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 862.929731] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 862.929982] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 862.930148] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 862.930649] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 862.930816] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 862.930965] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 862.931248] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.931378] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 862.931521] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 862.931691] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 862.931894] env[65385]: DEBUG nova.virt.hardware [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 862.933165] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f292ac-3ae3-4716-b55d-4a4d9fa38afa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.945654] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c3d501-73f1-4979-b830-4f57a53be4e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.973681] env[65385]: DEBUG nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 862.976482] env[65385]: INFO nova.compute.manager [-] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Took 1.48 seconds to deallocate network for instance. [ 863.012342] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453845, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070875} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.013054] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.013686] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a97ec4f-16fc-447d-8305-267365042b9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.038959] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.040426] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cd11266-101a-4b6d-b976-ffb8afbd9f68 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.064561] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 863.064561] env[65385]: value = "task-4453846" [ 863.064561] env[65385]: _type = "Task" [ 863.064561] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.073704] env[65385]: WARNING neutronclient.v2_0.client [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.073704] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.073886] env[65385]: WARNING openstack [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.083557] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.221s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.090161] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.888s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.091542] env[65385]: DEBUG nova.objects.instance [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lazy-loading 'resources' on Instance uuid b702ba27-bbc7-4030-a96d-30e1ed656a9b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.091757] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453846, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.103674] env[65385]: WARNING neutronclient.v2_0.client [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.104370] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.104719] env[65385]: WARNING openstack [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.118799] env[65385]: INFO nova.scheduler.client.report [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleted allocations for instance f25cc41e-560e-4cab-8bf0-dac47899d037 [ 863.147117] env[65385]: INFO nova.compute.manager [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] instance snapshotting [ 863.150591] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d99ced-11c9-42e6-980a-42402ee4ac16 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.174516] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.174777] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.174979] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.175187] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.175726] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.177774] env[65385]: INFO nova.compute.manager [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Terminating instance [ 863.179780] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4905189b-8a30-47f4-882a-02fba8f144ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.225580] env[65385]: DEBUG oslo_vmware.api [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453844, 'name': PowerOnVM_Task, 'duration_secs': 0.603917} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.225888] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.226107] env[65385]: INFO nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Took 10.60 seconds to spawn the instance on the hypervisor. [ 863.226563] env[65385]: DEBUG nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 863.228198] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcf5564-718d-4507-adef-bcf748fc8909 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.244494] env[65385]: DEBUG nova.compute.manager [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Received event network-vif-plugged-722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 863.244494] env[65385]: DEBUG oslo_concurrency.lockutils [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Acquiring lock "370b6aa8-2337-48e1-8102-611ce4494a55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.244494] env[65385]: DEBUG oslo_concurrency.lockutils [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Lock "370b6aa8-2337-48e1-8102-611ce4494a55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.244494] env[65385]: DEBUG oslo_concurrency.lockutils [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Lock "370b6aa8-2337-48e1-8102-611ce4494a55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.244494] env[65385]: DEBUG nova.compute.manager [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] No waiting events found dispatching network-vif-plugged-722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 863.244494] env[65385]: WARNING nova.compute.manager [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Received unexpected event network-vif-plugged-722cc380-0268-4488-ad6c-462f3af7b938 for instance with vm_state building and task_state spawning. [ 863.244494] env[65385]: DEBUG nova.compute.manager [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Received event network-changed-722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 863.244494] env[65385]: DEBUG nova.compute.manager [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Refreshing instance network info cache due to event network-changed-722cc380-0268-4488-ad6c-462f3af7b938. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 863.244494] env[65385]: DEBUG oslo_concurrency.lockutils [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Acquiring lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.244494] env[65385]: DEBUG oslo_concurrency.lockutils [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Acquired lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.244494] env[65385]: DEBUG nova.network.neutron [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Refreshing network info cache for port 722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 863.244494] env[65385]: WARNING oslo_messaging._drivers.amqpdriver [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 863.378792] env[65385]: INFO nova.compute.manager [-] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Took 1.66 seconds to deallocate network for instance. [ 863.383560] env[65385]: DEBUG nova.network.neutron [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Updated VIF entry in instance network info cache for port f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 863.384402] env[65385]: DEBUG nova.network.neutron [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Updating instance_info_cache with network_info: [{"id": "f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55", "address": "fa:16:3e:d5:05:ad", "network": {"id": "84b6a9d4-a34c-4839-8959-accb4a12f85b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1972478708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de072ae743844f6d9949acfe0fa6328c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6b500f6-8c", "ovs_interfaceid": "f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.442756] env[65385]: DEBUG nova.network.neutron [-] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.447849] env[65385]: DEBUG nova.compute.manager [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Received event network-vif-unplugged-d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 863.448034] env[65385]: DEBUG oslo_concurrency.lockutils [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Acquiring lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.448314] env[65385]: DEBUG oslo_concurrency.lockutils [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.448522] env[65385]: DEBUG oslo_concurrency.lockutils [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.448711] env[65385]: DEBUG nova.compute.manager [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] No waiting events found dispatching network-vif-unplugged-d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 863.449464] env[65385]: DEBUG nova.compute.manager [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Received event network-vif-unplugged-d15afa85-053b-49b7-a9df-eb592c2ba28d for instance with task_state deleting. {{(pid=65385) _process_instance_event /opt/stack/nova/nova/compute/manager.py:11597}} [ 863.449464] env[65385]: DEBUG nova.compute.manager [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Received event network-changed-d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 863.449464] env[65385]: DEBUG nova.compute.manager [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Refreshing instance network info cache due to event network-changed-d15afa85-053b-49b7-a9df-eb592c2ba28d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 863.449464] env[65385]: DEBUG oslo_concurrency.lockutils [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Acquiring lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.449705] env[65385]: DEBUG oslo_concurrency.lockutils [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Acquired lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.449748] env[65385]: DEBUG nova.network.neutron [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Refreshing network info cache for port d15afa85-053b-49b7-a9df-eb592c2ba28d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 863.482611] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.494748] env[65385]: DEBUG nova.network.neutron [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updated VIF entry in instance network info cache for port 7d96802a-1199-4a18-9a3d-f7b69a52590c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 863.495241] env[65385]: DEBUG nova.network.neutron [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updating instance_info_cache with network_info: [{"id": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "address": "fa:16:3e:42:06:97", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d96802a-11", "ovs_interfaceid": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.502692] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.578630] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.629138] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9b4790c3-56ed-4e0f-bde1-31949b8e4581 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "f25cc41e-560e-4cab-8bf0-dac47899d037" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.430s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.637106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.637361] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.637593] env[65385]: DEBUG nova.compute.manager [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Going to confirm migration 3 {{(pid=65385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 863.684386] env[65385]: DEBUG nova.network.neutron [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Successfully updated port: 3f2342b2-bc3c-489f-a61b-ffb18a20cce6 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 863.685769] env[65385]: DEBUG nova.compute.manager [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 863.685981] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.688950] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89d158a-6c79-4bdd-baf8-54c3544797cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.694608] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 863.695374] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3cff04e8-3c5f-42e3-a63e-5d5a367245c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.703602] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.707458] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2822418d-cb88-48fb-b85a-362e53d25b48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.708955] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 863.708955] env[65385]: value = "task-4453847" [ 863.708955] env[65385]: _type = "Task" [ 863.708955] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.710559] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 863.710671] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871070', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'name': 'volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '26a2d464-0aab-463f-8265-948a4f0ff188', 'attached_at': '', 'detached_at': '', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'serial': '33fd2b93-a90e-4837-9089-e9d5cab1e2df'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 863.713122] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c646be-b2d6-4ea5-8ec0-1ffc29863f64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.723780] env[65385]: DEBUG oslo_vmware.api [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 863.723780] env[65385]: value = "task-4453848" [ 863.723780] env[65385]: _type = "Task" [ 863.723780] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.759996] env[65385]: WARNING neutronclient.v2_0.client [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.759996] env[65385]: WARNING openstack [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.759996] env[65385]: WARNING openstack [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.765108] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453847, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.768668] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995183ce-3e87-422b-8fd9-ecb60ffc9e1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.780968] env[65385]: INFO nova.compute.manager [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Took 35.22 seconds to build instance. [ 863.783320] env[65385]: DEBUG oslo_vmware.api [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453848, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.821992] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df/volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.828029] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b905bec8-9de9-4270-b257-dcdda0c94deb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.850118] env[65385]: DEBUG oslo_vmware.api [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 863.850118] env[65385]: value = "task-4453849" [ 863.850118] env[65385]: _type = "Task" [ 863.850118] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.861414] env[65385]: DEBUG oslo_vmware.api [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.891622] env[65385]: DEBUG oslo_concurrency.lockutils [req-771c43f2-3bc8-4166-8ef2-a84415806831 req-ebbd1afe-bd18-4809-aaa6-9c28bdab4be8 service nova] Releasing lock "refresh_cache-9908d395-545d-4caf-9757-320d6253d61e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.893240] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.920663] env[65385]: WARNING openstack [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.921473] env[65385]: WARNING openstack [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.954277] env[65385]: INFO nova.compute.manager [-] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Took 1.48 seconds to deallocate network for instance. [ 863.954277] env[65385]: WARNING neutronclient.v2_0.client [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.954277] env[65385]: WARNING openstack [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.954277] env[65385]: WARNING openstack [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.004756] env[65385]: DEBUG oslo_concurrency.lockutils [req-0206eeee-360b-4e75-925f-f6ecc026fd74 req-f71743bf-d804-4ce6-a4a8-6bb8e00df622 service nova] Releasing lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.033901] env[65385]: WARNING neutronclient.v2_0.client [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.034959] env[65385]: WARNING openstack [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.035627] env[65385]: WARNING openstack [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.085611] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453846, 'name': ReconfigVM_Task, 'duration_secs': 0.777514} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.086273] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.087377] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-892b1fde-4a9f-4a62-bc09-fbc30a29e8ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.094735] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 864.094735] env[65385]: value = "task-4453850" [ 864.094735] env[65385]: _type = "Task" [ 864.094735] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.107283] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453850, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.129054] env[65385]: WARNING openstack [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.129823] env[65385]: WARNING openstack [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.146778] env[65385]: WARNING neutronclient.v2_0.client [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.165261] env[65385]: DEBUG nova.network.neutron [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Updated VIF entry in instance network info cache for port 722cc380-0268-4488-ad6c-462f3af7b938. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 864.165800] env[65385]: DEBUG nova.network.neutron [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Updating instance_info_cache with network_info: [{"id": "722cc380-0268-4488-ad6c-462f3af7b938", "address": "fa:16:3e:a9:5f:40", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap722cc380-02", "ovs_interfaceid": "722cc380-0268-4488-ad6c-462f3af7b938", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.188658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "refresh_cache-0b02fcd5-9fc7-4543-a754-d31fa0186981" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.188841] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquired lock "refresh_cache-0b02fcd5-9fc7-4543-a754-d31fa0186981" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.189067] env[65385]: DEBUG nova.network.neutron [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 864.209798] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d12166-75b3-495e-baed-60e1096b7501 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.228446] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3a849a-2794-493b-95b2-21f9a4037f05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.232122] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453847, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.233398] env[65385]: WARNING neutronclient.v2_0.client [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.234345] env[65385]: WARNING openstack [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.234464] env[65385]: WARNING openstack [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.253022] env[65385]: DEBUG oslo_vmware.api [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453848, 'name': PowerOffVM_Task, 'duration_secs': 0.268753} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.279357] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.279507] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.281699] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db0accf9-1074-4f45-8547-a2d73d509128 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.284262] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f271d295-6af9-446b-a232-c2654cfc6e28 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.289504] env[65385]: WARNING neutronclient.v2_0.client [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.289954] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.290103] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.290165] env[65385]: DEBUG nova.network.neutron [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 864.290350] env[65385]: DEBUG nova.objects.instance [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'info_cache' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.292008] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dc9b170d-7031-41f4-bd33-8cbe1d85e8c3 tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "9908d395-545d-4caf-9757-320d6253d61e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.738s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.299915] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9252b7a4-0646-4477-98ab-a221b280c090 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.322630] env[65385]: DEBUG nova.compute.provider_tree [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.365244] env[65385]: DEBUG oslo_vmware.api [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453849, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.369715] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.369897] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.369993] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleting the datastore file [datastore2] 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.370384] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95f9f0ee-2e14-4433-9e89-254a42fc46b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.379123] env[65385]: DEBUG oslo_vmware.api [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 864.379123] env[65385]: value = "task-4453852" [ 864.379123] env[65385]: _type = "Task" [ 864.379123] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.388726] env[65385]: DEBUG oslo_vmware.api [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.393091] env[65385]: DEBUG nova.network.neutron [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Updated VIF entry in instance network info cache for port d15afa85-053b-49b7-a9df-eb592c2ba28d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 864.393582] env[65385]: DEBUG nova.network.neutron [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Updating instance_info_cache with network_info: [{"id": "d15afa85-053b-49b7-a9df-eb592c2ba28d", "address": "fa:16:3e:f3:7a:17", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": null, "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd15afa85-05", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.475623] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.610047] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453850, 'name': Rename_Task, 'duration_secs': 0.17549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.610593] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.610985] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8de82b3-2235-4822-a464-a4d6242494c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.622142] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 864.622142] env[65385]: value = "task-4453853" [ 864.622142] env[65385]: _type = "Task" [ 864.622142] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.633397] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.669321] env[65385]: DEBUG oslo_concurrency.lockutils [req-3523931a-8bde-402b-b6dc-6b086af16183 req-78702664-1c5b-45a3-9fd0-c6e2addb043a service nova] Releasing lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.692617] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.693090] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.727308] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453847, 'name': CreateSnapshot_Task, 'duration_secs': 0.810974} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.727605] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 864.728568] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd5c9a4-3c05-423e-857d-2e8e6c79c8d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.732963] env[65385]: DEBUG nova.network.neutron [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 864.759729] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.760202] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.827242] env[65385]: DEBUG nova.scheduler.client.report [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 864.845281] env[65385]: WARNING neutronclient.v2_0.client [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.845881] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.846271] env[65385]: WARNING openstack [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.866354] env[65385]: DEBUG oslo_vmware.api [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453849, 'name': ReconfigVM_Task, 'duration_secs': 0.595981} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.866747] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Reconfigured VM instance instance-00000037 to attach disk [datastore2] volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df/volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.872411] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b0f5c48-8009-412c-b347-9eb5e592aa35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.893731] env[65385]: DEBUG oslo_vmware.api [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453852, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.293441} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.899710] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.899918] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.900252] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.900366] env[65385]: INFO nova.compute.manager [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Took 1.21 seconds to destroy the instance on the hypervisor. [ 864.900615] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 864.901045] env[65385]: DEBUG oslo_concurrency.lockutils [req-c664f3ff-d898-4577-b928-24497d381f40 req-08e5e516-007b-46b4-822b-831fb9bfc1e7 service nova] Releasing lock "refresh_cache-b1cca536-e1d9-43f8-8a5c-be3981c106e7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.901912] env[65385]: DEBUG oslo_vmware.api [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 864.901912] env[65385]: value = "task-4453854" [ 864.901912] env[65385]: _type = "Task" [ 864.901912] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.902347] env[65385]: DEBUG nova.compute.manager [-] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 864.902447] env[65385]: DEBUG nova.network.neutron [-] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 864.902691] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.903254] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.903504] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.920129] env[65385]: DEBUG oslo_vmware.api [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.954976] env[65385]: DEBUG nova.network.neutron [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Updating instance_info_cache with network_info: [{"id": "3f2342b2-bc3c-489f-a61b-ffb18a20cce6", "address": "fa:16:3e:a2:42:52", "network": {"id": "44d252ad-98ef-4354-a13b-331d4950e802", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1117895196-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc464c9d4e49499ba42b61ce293ac4fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f2342b2-bc", "ovs_interfaceid": "3f2342b2-bc3c-489f-a61b-ffb18a20cce6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.004546] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.047040] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "65f9bc55-7bab-43b0-a974-eb5080389b7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.047251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "65f9bc55-7bab-43b0-a974-eb5080389b7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.047467] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "65f9bc55-7bab-43b0-a974-eb5080389b7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.047649] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "65f9bc55-7bab-43b0-a974-eb5080389b7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.047818] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "65f9bc55-7bab-43b0-a974-eb5080389b7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.050564] env[65385]: INFO nova.compute.manager [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Terminating instance [ 865.132679] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453853, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.252894] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 865.253634] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-da3cab26-65b8-414b-98fb-21d7cd495681 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.263057] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 865.263057] env[65385]: value = "task-4453855" [ 865.263057] env[65385]: _type = "Task" [ 865.263057] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.274906] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453855, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.301895] env[65385]: WARNING neutronclient.v2_0.client [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.302927] env[65385]: WARNING openstack [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.303977] env[65385]: WARNING openstack [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.316451] env[65385]: DEBUG nova.compute.manager [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 865.320432] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc6292a-fc97-4790-919b-cc3a824f8a0d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.329871] env[65385]: DEBUG nova.compute.manager [req-cff4bdeb-0c5b-423e-ab18-e3cedcae3be0 req-9641e78a-54d6-405d-b7cf-3303a44c11b6 service nova] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Received event network-vif-deleted-ad95d76f-e023-47b2-a936-e7ed11779276 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.330261] env[65385]: DEBUG nova.compute.manager [req-cff4bdeb-0c5b-423e-ab18-e3cedcae3be0 req-9641e78a-54d6-405d-b7cf-3303a44c11b6 service nova] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Received event network-vif-deleted-f61ecab0-4449-47db-8bfb-6ff1499f6d71 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.330547] env[65385]: DEBUG nova.compute.manager [req-cff4bdeb-0c5b-423e-ab18-e3cedcae3be0 req-9641e78a-54d6-405d-b7cf-3303a44c11b6 service nova] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Received event network-vif-deleted-d59b902f-9fb1-4d56-96f9-6301024412fa {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.330815] env[65385]: DEBUG nova.compute.manager [req-cff4bdeb-0c5b-423e-ab18-e3cedcae3be0 req-9641e78a-54d6-405d-b7cf-3303a44c11b6 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Received event network-vif-deleted-97b8305b-a557-46c0-9f9a-f45d7ccd6e3a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.331104] env[65385]: INFO nova.compute.manager [req-cff4bdeb-0c5b-423e-ab18-e3cedcae3be0 req-9641e78a-54d6-405d-b7cf-3303a44c11b6 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Neutron deleted interface 97b8305b-a557-46c0-9f9a-f45d7ccd6e3a; detaching it from the instance and deleting it from the info cache [ 865.331368] env[65385]: DEBUG nova.network.neutron [req-cff4bdeb-0c5b-423e-ab18-e3cedcae3be0 req-9641e78a-54d6-405d-b7cf-3303a44c11b6 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.334504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.244s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.339962] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.104s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.340387] env[65385]: DEBUG nova.objects.instance [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lazy-loading 'resources' on Instance uuid f521e37d-aa86-4533-9fbe-7d040ed90e53 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.381489] env[65385]: INFO nova.scheduler.client.report [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleted allocations for instance b702ba27-bbc7-4030-a96d-30e1ed656a9b [ 865.429598] env[65385]: DEBUG oslo_vmware.api [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453854, 'name': ReconfigVM_Task, 'duration_secs': 0.286549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.430441] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871070', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'name': 'volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '26a2d464-0aab-463f-8265-948a4f0ff188', 'attached_at': '', 'detached_at': '', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'serial': '33fd2b93-a90e-4837-9089-e9d5cab1e2df'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 865.458284] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Releasing lock "refresh_cache-0b02fcd5-9fc7-4543-a754-d31fa0186981" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.458715] env[65385]: DEBUG nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Instance network_info: |[{"id": "3f2342b2-bc3c-489f-a61b-ffb18a20cce6", "address": "fa:16:3e:a2:42:52", "network": {"id": "44d252ad-98ef-4354-a13b-331d4950e802", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1117895196-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc464c9d4e49499ba42b61ce293ac4fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f2342b2-bc", "ovs_interfaceid": "3f2342b2-bc3c-489f-a61b-ffb18a20cce6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 865.459576] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:42:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496ac502-bfc4-4324-8332-cac473eb7cc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f2342b2-bc3c-489f-a61b-ffb18a20cce6', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.467302] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Creating folder: Project (bc464c9d4e49499ba42b61ce293ac4fb). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.468105] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5307f017-a53c-4a5c-a96a-843513b59ac0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.483037] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Created folder: Project (bc464c9d4e49499ba42b61ce293ac4fb) in parent group-v870881. [ 865.483037] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Creating folder: Instances. Parent ref: group-v871074. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.483037] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0328d6f7-7868-4d3d-abdb-feaf616aec35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.496445] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Created folder: Instances in parent group-v871074. [ 865.496728] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 865.497365] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.497611] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d64a6bd-d0d2-4624-a525-d3a9577b8feb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.525155] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.525155] env[65385]: value = "task-4453858" [ 865.525155] env[65385]: _type = "Task" [ 865.525155] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.538459] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453858, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.552808] env[65385]: WARNING openstack [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.553511] env[65385]: WARNING openstack [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.563637] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "refresh_cache-65f9bc55-7bab-43b0-a974-eb5080389b7d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.563983] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquired lock "refresh_cache-65f9bc55-7bab-43b0-a974-eb5080389b7d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.564278] env[65385]: DEBUG nova.network.neutron [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 865.638453] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453853, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.676255] env[65385]: WARNING neutronclient.v2_0.client [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.676947] env[65385]: WARNING openstack [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.677331] env[65385]: WARNING openstack [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.734509] env[65385]: DEBUG nova.network.neutron [-] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.802550] env[65385]: DEBUG nova.compute.manager [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Received event network-vif-plugged-3f2342b2-bc3c-489f-a61b-ffb18a20cce6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.802550] env[65385]: DEBUG oslo_concurrency.lockutils [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Acquiring lock "0b02fcd5-9fc7-4543-a754-d31fa0186981-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.802550] env[65385]: DEBUG oslo_concurrency.lockutils [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.802550] env[65385]: DEBUG oslo_concurrency.lockutils [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.802550] env[65385]: DEBUG nova.compute.manager [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] No waiting events found dispatching network-vif-plugged-3f2342b2-bc3c-489f-a61b-ffb18a20cce6 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 865.802550] env[65385]: WARNING nova.compute.manager [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Received unexpected event network-vif-plugged-3f2342b2-bc3c-489f-a61b-ffb18a20cce6 for instance with vm_state building and task_state spawning. [ 865.802550] env[65385]: DEBUG nova.compute.manager [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Received event network-changed-3f2342b2-bc3c-489f-a61b-ffb18a20cce6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.802550] env[65385]: DEBUG nova.compute.manager [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Refreshing instance network info cache due to event network-changed-3f2342b2-bc3c-489f-a61b-ffb18a20cce6. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 865.802550] env[65385]: DEBUG oslo_concurrency.lockutils [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Acquiring lock "refresh_cache-0b02fcd5-9fc7-4543-a754-d31fa0186981" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.802550] env[65385]: DEBUG oslo_concurrency.lockutils [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Acquired lock "refresh_cache-0b02fcd5-9fc7-4543-a754-d31fa0186981" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.802550] env[65385]: DEBUG nova.network.neutron [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Refreshing network info cache for port 3f2342b2-bc3c-489f-a61b-ffb18a20cce6 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 865.802550] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453855, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.826656] env[65385]: DEBUG nova.network.neutron [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.842662] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36a21af7-a3cb-4fce-90eb-ca2504d486ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.854718] env[65385]: INFO nova.compute.manager [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] instance snapshotting [ 865.857961] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb5e0ac-f997-48f9-9b33-6631a8ad7f70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.869136] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ebeee1-6d69-49cb-992a-32b67146a79c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.904028] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eae4339-e9cf-494c-8b3c-3ab2b48a4a80 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.908047] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e7a53b51-4817-4f59-82d7-689905a43bd8 tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "b702ba27-bbc7-4030-a96d-30e1ed656a9b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.485s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.942205] env[65385]: DEBUG nova.compute.manager [req-cff4bdeb-0c5b-423e-ab18-e3cedcae3be0 req-9641e78a-54d6-405d-b7cf-3303a44c11b6 service nova] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Detach interface failed, port_id=97b8305b-a557-46c0-9f9a-f45d7ccd6e3a, reason: Instance 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 866.041950] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453858, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.067042] env[65385]: WARNING neutronclient.v2_0.client [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.068468] env[65385]: WARNING openstack [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.068468] env[65385]: WARNING openstack [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.095312] env[65385]: DEBUG nova.network.neutron [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 866.133796] env[65385]: DEBUG oslo_vmware.api [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4453853, 'name': PowerOnVM_Task, 'duration_secs': 1.492851} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.134103] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.134309] env[65385]: INFO nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Took 8.10 seconds to spawn the instance on the hypervisor. [ 866.134685] env[65385]: DEBUG nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 866.135359] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef6e738-a9cc-40ef-9080-c5fa9f7d9e3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.148518] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "9908d395-545d-4caf-9757-320d6253d61e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.148788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "9908d395-545d-4caf-9757-320d6253d61e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.149373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "9908d395-545d-4caf-9757-320d6253d61e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.149373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "9908d395-545d-4caf-9757-320d6253d61e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.149516] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "9908d395-545d-4caf-9757-320d6253d61e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 866.151547] env[65385]: INFO nova.compute.manager [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Terminating instance [ 866.169260] env[65385]: DEBUG nova.network.neutron [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.239840] env[65385]: INFO nova.compute.manager [-] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Took 1.34 seconds to deallocate network for instance. [ 866.265204] env[65385]: WARNING neutronclient.v2_0.client [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.265871] env[65385]: WARNING openstack [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.266282] env[65385]: WARNING openstack [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.288525] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453855, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.331303] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.332631] env[65385]: DEBUG nova.objects.instance [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'migration_context' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.347437] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2533ccd0-cb5d-4aa1-b3f6-81204bbeb786 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.356776] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1813eb-23c4-4b27-86fa-97c47c0033e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.395788] env[65385]: WARNING openstack [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.396117] env[65385]: WARNING openstack [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.403167] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee595ec-eefd-49b3-a196-570da1f25e32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.412367] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969c3b9a-9649-4074-a9d5-13486307cb77 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.430847] env[65385]: DEBUG nova.compute.provider_tree [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.450314] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 866.451379] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-29965683-5a08-4edf-9fa3-ce64004ecc0c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.465024] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 866.465024] env[65385]: value = "task-4453859" [ 866.465024] env[65385]: _type = "Task" [ 866.465024] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.476318] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453859, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.486831] env[65385]: WARNING neutronclient.v2_0.client [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.487682] env[65385]: WARNING openstack [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.488127] env[65385]: WARNING openstack [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.506033] env[65385]: DEBUG nova.objects.instance [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'flavor' on Instance uuid 26a2d464-0aab-463f-8265-948a4f0ff188 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.540109] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453858, 'name': CreateVM_Task, 'duration_secs': 0.615383} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.540286] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.540782] env[65385]: WARNING neutronclient.v2_0.client [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.541198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.541362] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.541665] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 866.542190] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8f2fbe8-2df9-48a1-ae08-fccc54bb8b64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.552501] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 866.552501] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5222d9b1-360a-b7a4-3b44-abd3e936a3f1" [ 866.552501] env[65385]: _type = "Task" [ 866.552501] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.562729] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5222d9b1-360a-b7a4-3b44-abd3e936a3f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.595368] env[65385]: DEBUG nova.network.neutron [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Updated VIF entry in instance network info cache for port 3f2342b2-bc3c-489f-a61b-ffb18a20cce6. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 866.595774] env[65385]: DEBUG nova.network.neutron [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Updating instance_info_cache with network_info: [{"id": "3f2342b2-bc3c-489f-a61b-ffb18a20cce6", "address": "fa:16:3e:a2:42:52", "network": {"id": "44d252ad-98ef-4354-a13b-331d4950e802", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1117895196-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc464c9d4e49499ba42b61ce293ac4fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f2342b2-bc", "ovs_interfaceid": "3f2342b2-bc3c-489f-a61b-ffb18a20cce6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.659827] env[65385]: INFO nova.compute.manager [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Took 34.81 seconds to build instance. [ 866.661339] env[65385]: DEBUG nova.compute.manager [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 866.661646] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.663042] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac9d7de-f59a-4982-bd8f-4a804452f5f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.673111] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.673761] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f07073e6-dc3d-4684-8ba9-5dd61c50759f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.676419] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Releasing lock "refresh_cache-65f9bc55-7bab-43b0-a974-eb5080389b7d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.676823] env[65385]: DEBUG nova.compute.manager [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 866.677017] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.678124] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a31e0fa-9576-49a2-831c-746ae5d37252 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.688576] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.689836] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b9c5187-30ed-44fd-add5-5289d7c06748 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.691803] env[65385]: DEBUG oslo_vmware.api [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 866.691803] env[65385]: value = "task-4453860" [ 866.691803] env[65385]: _type = "Task" [ 866.691803] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.698697] env[65385]: DEBUG oslo_vmware.api [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 866.698697] env[65385]: value = "task-4453861" [ 866.698697] env[65385]: _type = "Task" [ 866.698697] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.707213] env[65385]: DEBUG oslo_vmware.api [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.717324] env[65385]: DEBUG oslo_vmware.api [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.747732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.785080] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453855, 'name': CloneVM_Task, 'duration_secs': 1.40774} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.785487] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Created linked-clone VM from snapshot [ 866.786336] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820ac0d0-0074-4ae9-8453-0e73650c1a35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.795892] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Uploading image c3f38eb9-8503-49ea-a6f5-d7759c1ac49c {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 866.829515] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 866.829515] env[65385]: value = "vm-871073" [ 866.829515] env[65385]: _type = "VirtualMachine" [ 866.829515] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 866.829854] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7208efce-21fb-4950-bc18-518c8bdb0a85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.834989] env[65385]: DEBUG nova.objects.base [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Object Instance<89f3c815-8671-47ce-9e74-bf6e652bb3c3> lazy-loaded attributes: info_cache,migration_context {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 866.836016] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5acea1-49a6-4ca6-aeb7-1020e46b691d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.840412] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lease: (returnval){ [ 866.840412] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52555e5d-719f-5453-ec68-752b2f91eddb" [ 866.840412] env[65385]: _type = "HttpNfcLease" [ 866.840412] env[65385]: } obtained for exporting VM: (result){ [ 866.840412] env[65385]: value = "vm-871073" [ 866.840412] env[65385]: _type = "VirtualMachine" [ 866.840412] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 866.840741] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the lease: (returnval){ [ 866.840741] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52555e5d-719f-5453-ec68-752b2f91eddb" [ 866.840741] env[65385]: _type = "HttpNfcLease" [ 866.840741] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 866.860959] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a7a0d96-439e-4eb4-84b1-868898cb1afd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.865344] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 866.865344] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52555e5d-719f-5453-ec68-752b2f91eddb" [ 866.865344] env[65385]: _type = "HttpNfcLease" [ 866.865344] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 866.866313] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 866.866313] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52555e5d-719f-5453-ec68-752b2f91eddb" [ 866.866313] env[65385]: _type = "HttpNfcLease" [ 866.866313] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 866.867182] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821f2a4e-2a5e-4a9d-be06-259222cd0417 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.872807] env[65385]: DEBUG oslo_vmware.api [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 866.872807] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e8fac-82bb-f15d-1e85-8981cc296ebe" [ 866.872807] env[65385]: _type = "Task" [ 866.872807] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.881950] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52385975-7571-e4eb-5b74-9c3277a0c606/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 866.882165] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52385975-7571-e4eb-5b74-9c3277a0c606/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 866.944057] env[65385]: DEBUG nova.scheduler.client.report [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 866.952925] env[65385]: DEBUG oslo_vmware.api [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e8fac-82bb-f15d-1e85-8981cc296ebe, 'name': SearchDatastore_Task, 'duration_secs': 0.011292} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.953781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.974811] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453859, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.016504] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-54e56bb5-014b-44df-99d4-67db4089a367 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.018076] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1dd634e4-a9f2-4671-b77b-cd5710fbb2be tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.975s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.067910] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5222d9b1-360a-b7a4-3b44-abd3e936a3f1, 'name': SearchDatastore_Task, 'duration_secs': 0.015236} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.068195] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.068418] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.068647] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.068788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 867.068956] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.069273] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbba12ed-fc9e-4dc9-a7f4-11077b89913a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.083037] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.083183] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.083929] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffbfa734-4c40-48ad-a88d-440398ca874a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.089942] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 867.089942] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524bc543-a558-850e-023e-23b7728f5d1c" [ 867.089942] env[65385]: _type = "Task" [ 867.089942] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.098922] env[65385]: DEBUG oslo_concurrency.lockutils [req-7f902bf1-7f93-4b19-af89-61925eac1518 req-25d217b1-161d-4fa2-87ff-b7b6eec52fee service nova] Releasing lock "refresh_cache-0b02fcd5-9fc7-4543-a754-d31fa0186981" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.099398] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524bc543-a558-850e-023e-23b7728f5d1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.165102] env[65385]: DEBUG oslo_concurrency.lockutils [None req-189ed8b6-3523-4d01-8bc6-148070a8f6da tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "370b6aa8-2337-48e1-8102-611ce4494a55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.331s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.203877] env[65385]: DEBUG oslo_vmware.api [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453860, 'name': PowerOffVM_Task, 'duration_secs': 0.308858} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.206990] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.206990] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.207152] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40e7f561-a673-4666-b9a8-ad29d1795f9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.216070] env[65385]: DEBUG oslo_vmware.api [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453861, 'name': PowerOffVM_Task, 'duration_secs': 0.241521} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.216070] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.216070] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.216070] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f04def1-f72e-43e6-bba0-b7fa9284ec55 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.246429] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.246905] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.247231] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleting the datastore file [datastore1] 65f9bc55-7bab-43b0-a974-eb5080389b7d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.247564] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef7310e3-fa40-44c6-99b3-5adaba2785f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.255561] env[65385]: DEBUG oslo_vmware.api [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for the task: (returnval){ [ 867.255561] env[65385]: value = "task-4453865" [ 867.255561] env[65385]: _type = "Task" [ 867.255561] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.267849] env[65385]: DEBUG oslo_vmware.api [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453865, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.292018] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.292387] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.292651] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Deleting the datastore file [datastore2] 9908d395-545d-4caf-9757-320d6253d61e {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.293083] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ab2c721-195a-421c-b98b-0831d7b7b395 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.302381] env[65385]: DEBUG oslo_vmware.api [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for the task: (returnval){ [ 867.302381] env[65385]: value = "task-4453866" [ 867.302381] env[65385]: _type = "Task" [ 867.302381] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.314441] env[65385]: DEBUG oslo_vmware.api [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.449930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.110s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.456029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.097s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.457506] env[65385]: INFO nova.compute.claims [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.477988] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453859, 'name': CreateSnapshot_Task, 'duration_secs': 1.004784} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.478597] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 867.479829] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fce1b2c-554c-419d-b6d9-1c4139cf5da9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.510156] env[65385]: INFO nova.scheduler.client.report [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Deleted allocations for instance f521e37d-aa86-4533-9fbe-7d040ed90e53 [ 867.603130] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524bc543-a558-850e-023e-23b7728f5d1c, 'name': SearchDatastore_Task, 'duration_secs': 0.033611} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.603835] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-613523fa-d4f0-4615-8e24-82a355060b17 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.612905] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 867.612905] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524fc63b-acad-ae1f-f108-09e7fdf8ab96" [ 867.612905] env[65385]: _type = "Task" [ 867.612905] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.626054] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524fc63b-acad-ae1f-f108-09e7fdf8ab96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.767083] env[65385]: DEBUG oslo_vmware.api [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Task: {'id': task-4453865, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179636} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.770343] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.770343] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.770343] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.770343] env[65385]: INFO nova.compute.manager [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Took 1.09 seconds to destroy the instance on the hypervisor. [ 867.770343] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 867.770343] env[65385]: DEBUG nova.compute.manager [-] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 867.770343] env[65385]: DEBUG nova.network.neutron [-] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 867.770343] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.770343] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.770343] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.814345] env[65385]: DEBUG oslo_vmware.api [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Task: {'id': task-4453866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27604} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.814652] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.814839] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.815042] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.815605] env[65385]: INFO nova.compute.manager [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 867.815605] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 867.815774] env[65385]: DEBUG nova.compute.manager [-] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 867.815938] env[65385]: DEBUG nova.network.neutron [-] [instance: 9908d395-545d-4caf-9757-320d6253d61e] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 867.816269] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.817055] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.817515] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.004025] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 868.004025] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ed62ee03-87ee-4af1-b407-b171c22d01e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.025067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-764daa09-ce69-4e9d-bf7b-e0c9afd1a3df tempest-ListServersNegativeTestJSON-832628132 tempest-ListServersNegativeTestJSON-832628132-project-member] Lock "f521e37d-aa86-4533-9fbe-7d040ed90e53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.288s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.028300] env[65385]: DEBUG nova.network.neutron [-] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 868.030078] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.032147] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 868.032147] env[65385]: value = "task-4453867" [ 868.032147] env[65385]: _type = "Task" [ 868.032147] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.048509] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453867, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.066828] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.128260] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524fc63b-acad-ae1f-f108-09e7fdf8ab96, 'name': SearchDatastore_Task, 'duration_secs': 0.017526} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.129075] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.129075] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 0b02fcd5-9fc7-4543-a754-d31fa0186981/0b02fcd5-9fc7-4543-a754-d31fa0186981.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.129558] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f41d902-1871-4d2d-a11c-510f63f064b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.141840] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 868.141840] env[65385]: value = "task-4453868" [ 868.141840] env[65385]: _type = "Task" [ 868.141840] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.155393] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453868, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.536190] env[65385]: DEBUG nova.network.neutron [-] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 868.554687] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453867, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.666779] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453868, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.044902] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300f98e2-8811-4a56-8dd1-d5728bef1bcd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.053712] env[65385]: INFO nova.compute.manager [-] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Took 1.28 seconds to deallocate network for instance. [ 869.054294] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453867, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.066301] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78605f21-32a0-4254-bede-2f4d1c090203 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.073675] env[65385]: DEBUG nova.network.neutron [-] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 869.114621] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c68eaf-62a6-42a2-a059-935fd50f6971 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.124476] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b25d070-4364-4306-a9e2-0abaf6499e6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.143025] env[65385]: DEBUG nova.compute.provider_tree [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.156763] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453868, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649425} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.158099] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 0b02fcd5-9fc7-4543-a754-d31fa0186981/0b02fcd5-9fc7-4543-a754-d31fa0186981.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.158323] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.158589] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c334c66-2c93-4d2b-80a4-6cd2c7dba784 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.167269] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 869.167269] env[65385]: value = "task-4453869" [ 869.167269] env[65385]: _type = "Task" [ 869.167269] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.180069] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453869, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.548313] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453867, 'name': CloneVM_Task, 'duration_secs': 1.453764} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.548756] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Created linked-clone VM from snapshot [ 869.551381] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7756b0bf-5fe0-41f9-9027-be05e664dd06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.558729] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Uploading image 57fab5fe-ac7e-4d91-94ca-0285c2957e33 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 869.566701] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.577484] env[65385]: INFO nova.compute.manager [-] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Took 1.76 seconds to deallocate network for instance. [ 869.593824] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 869.593824] env[65385]: value = "vm-871078" [ 869.593824] env[65385]: _type = "VirtualMachine" [ 869.593824] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 869.598030] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bec03f61-7679-462c-b4d6-ecd993831bd5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.604707] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease: (returnval){ [ 869.604707] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e3140c-cc03-3706-03e4-c5fb947029cd" [ 869.604707] env[65385]: _type = "HttpNfcLease" [ 869.604707] env[65385]: } obtained for exporting VM: (result){ [ 869.604707] env[65385]: value = "vm-871078" [ 869.604707] env[65385]: _type = "VirtualMachine" [ 869.604707] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 869.605064] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the lease: (returnval){ [ 869.605064] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e3140c-cc03-3706-03e4-c5fb947029cd" [ 869.605064] env[65385]: _type = "HttpNfcLease" [ 869.605064] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 869.615875] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 869.615875] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e3140c-cc03-3706-03e4-c5fb947029cd" [ 869.615875] env[65385]: _type = "HttpNfcLease" [ 869.615875] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 869.648072] env[65385]: DEBUG nova.scheduler.client.report [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.677776] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453869, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084492} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.678073] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.678904] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d327dda7-550e-4ba1-b3b5-ba4f5055e9b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.703242] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 0b02fcd5-9fc7-4543-a754-d31fa0186981/0b02fcd5-9fc7-4543-a754-d31fa0186981.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.703977] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac29e56f-4cb8-4f52-b221-f4c2447f7038 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.724988] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 869.724988] env[65385]: value = "task-4453871" [ 869.724988] env[65385]: _type = "Task" [ 869.724988] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.734151] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453871, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.091319] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.118910] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 870.118910] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e3140c-cc03-3706-03e4-c5fb947029cd" [ 870.118910] env[65385]: _type = "HttpNfcLease" [ 870.118910] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 870.118910] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 870.118910] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e3140c-cc03-3706-03e4-c5fb947029cd" [ 870.118910] env[65385]: _type = "HttpNfcLease" [ 870.118910] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 870.118910] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434c97f9-acea-40ae-ab30-b8b75d87765d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.137966] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5292678e-0fac-a45b-ab8e-472e49b928a6/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 870.138357] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5292678e-0fac-a45b-ab8e-472e49b928a6/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 870.200390] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.746s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.201036] env[65385]: DEBUG nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 870.205726] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.077s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 870.205999] env[65385]: DEBUG nova.objects.instance [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lazy-loading 'resources' on Instance uuid 90c36f67-0c25-4487-a5e4-89827290953b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.247591] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.253252] env[65385]: DEBUG nova.compute.manager [req-6b7e1c2a-e3bb-4a5a-a911-2f2f8f26377a req-94980267-0db4-48c8-9407-74a47c036e17 service nova] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Received event network-vif-deleted-f6b500f6-8c2f-457c-b5e5-5b2adc9fbb55 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 870.255430] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-578dc3a9-8947-437d-979e-eca2f156130c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.707829] env[65385]: DEBUG nova.compute.utils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 870.709650] env[65385]: DEBUG nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 870.709967] env[65385]: DEBUG nova.network.neutron [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 870.710489] env[65385]: WARNING neutronclient.v2_0.client [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.710938] env[65385]: WARNING neutronclient.v2_0.client [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.711578] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.711918] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.738237] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453871, 'name': ReconfigVM_Task, 'duration_secs': 0.841713} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.738571] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 0b02fcd5-9fc7-4543-a754-d31fa0186981/0b02fcd5-9fc7-4543-a754-d31fa0186981.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.739490] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c37bbd8-3cff-4b82-8e56-17492a29a788 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.749693] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 870.749693] env[65385]: value = "task-4453872" [ 870.749693] env[65385]: _type = "Task" [ 870.749693] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.764976] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453872, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.054816] env[65385]: DEBUG nova.policy [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bbc524f6b2e64dd0af9d06525190299e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99f640b7d5594400bbd69c05713089d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 871.222584] env[65385]: DEBUG nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 871.230818] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09933116-b654-4141-8eaa-128018fe2747 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.241344] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d54f3c3-2aba-4454-ad26-6015cf8cf070 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.298102] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fd3d76-da9d-4162-a3fd-e59b65ca776f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.310867] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453872, 'name': Rename_Task, 'duration_secs': 0.281448} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.313729] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.314092] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d60dcf44-ec11-4ef5-b5f7-7ca33d7bde40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.317330] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444ea746-6e61-433f-a32f-aa9cbdb2f0ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.334195] env[65385]: DEBUG nova.compute.provider_tree [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.337880] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 871.337880] env[65385]: value = "task-4453873" [ 871.337880] env[65385]: _type = "Task" [ 871.337880] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.350411] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453873, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.453636] env[65385]: DEBUG nova.network.neutron [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Successfully created port: 2e9cf308-c94b-4a76-9258-93c5d7563146 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 871.837632] env[65385]: DEBUG nova.scheduler.client.report [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.852393] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453873, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.236408] env[65385]: DEBUG nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 872.248219] env[65385]: DEBUG nova.compute.manager [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Received event network-changed-722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 872.248334] env[65385]: DEBUG nova.compute.manager [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Refreshing instance network info cache due to event network-changed-722cc380-0268-4488-ad6c-462f3af7b938. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 872.248706] env[65385]: DEBUG oslo_concurrency.lockutils [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Acquiring lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.248927] env[65385]: DEBUG oslo_concurrency.lockutils [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Acquired lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 872.249182] env[65385]: DEBUG nova.network.neutron [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Refreshing network info cache for port 722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 872.259357] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "4b0b8859-b63e-4740-bf67-5733c75e1c70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.259658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 872.347270] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.141s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.349962] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.824s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 872.351561] env[65385]: INFO nova.compute.claims [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.365336] env[65385]: DEBUG oslo_vmware.api [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453873, 'name': PowerOnVM_Task, 'duration_secs': 0.689182} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.365636] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.365842] env[65385]: INFO nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Took 9.47 seconds to spawn the instance on the hypervisor. [ 872.366145] env[65385]: DEBUG nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 872.366926] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a29873-9ee2-4112-9739-7983d2fc8e95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.378460] env[65385]: INFO nova.scheduler.client.report [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Deleted allocations for instance 90c36f67-0c25-4487-a5e4-89827290953b [ 872.760079] env[65385]: WARNING neutronclient.v2_0.client [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 872.760822] env[65385]: WARNING openstack [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 872.761204] env[65385]: WARNING openstack [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.770042] env[65385]: DEBUG nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 872.898126] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c808693a-1529-4341-b17d-4586e129c51e tempest-ListImageFiltersTestJSON-1717330648 tempest-ListImageFiltersTestJSON-1717330648-project-member] Lock "90c36f67-0c25-4487-a5e4-89827290953b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.476s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.905072] env[65385]: INFO nova.compute.manager [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Took 36.23 seconds to build instance. [ 873.300220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.374385] env[65385]: DEBUG nova.network.neutron [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Successfully updated port: 2e9cf308-c94b-4a76-9258-93c5d7563146 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 873.407691] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06620f30-a691-4acf-8b90-1bde0c1944d1 tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.764s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.423900] env[65385]: WARNING openstack [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.424462] env[65385]: WARNING openstack [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.518865] env[65385]: WARNING neutronclient.v2_0.client [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 873.519689] env[65385]: WARNING openstack [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.520082] env[65385]: WARNING openstack [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.633522] env[65385]: DEBUG nova.network.neutron [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Updated VIF entry in instance network info cache for port 722cc380-0268-4488-ad6c-462f3af7b938. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 873.633697] env[65385]: DEBUG nova.network.neutron [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Updating instance_info_cache with network_info: [{"id": "722cc380-0268-4488-ad6c-462f3af7b938", "address": "fa:16:3e:a9:5f:40", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap722cc380-02", "ovs_interfaceid": "722cc380-0268-4488-ad6c-462f3af7b938", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 873.820055] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11ada58-82f6-486b-94e3-d0c65bb2b122 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.828515] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da68c68-4f71-4825-8f63-a7d4f5d9e714 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.860656] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d43fa0-75af-4234-9e0f-0da2a2b675a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.869893] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604a1db1-e1d3-4b9f-8935-f8a1c1322c30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.887409] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "refresh_cache-e02e3848-1b1e-426b-bc0f-24c8e232a9db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.887591] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "refresh_cache-e02e3848-1b1e-426b-bc0f-24c8e232a9db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.887770] env[65385]: DEBUG nova.network.neutron [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 873.889183] env[65385]: DEBUG nova.compute.provider_tree [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.137606] env[65385]: DEBUG oslo_concurrency.lockutils [req-2b988711-a12a-4e0c-9aa2-4403687054f5 req-4604bdf3-7540-4bab-8a51-9c010d4808fa service nova] Releasing lock "refresh_cache-370b6aa8-2337-48e1-8102-611ce4494a55" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.391645] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.392086] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.399945] env[65385]: DEBUG nova.scheduler.client.report [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 874.446649] env[65385]: DEBUG nova.network.neutron [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 874.468784] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.469210] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.538512] env[65385]: WARNING neutronclient.v2_0.client [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.539263] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.539721] env[65385]: WARNING openstack [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.639099] env[65385]: DEBUG nova.network.neutron [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Updating instance_info_cache with network_info: [{"id": "2e9cf308-c94b-4a76-9258-93c5d7563146", "address": "fa:16:3e:d1:6b:5e", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e9cf308-c9", "ovs_interfaceid": "2e9cf308-c94b-4a76-9258-93c5d7563146", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 874.905688] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.906368] env[65385]: DEBUG nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 874.909717] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.947s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.912059] env[65385]: INFO nova.compute.claims [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.142243] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "refresh_cache-e02e3848-1b1e-426b-bc0f-24c8e232a9db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.142766] env[65385]: DEBUG nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Instance network_info: |[{"id": "2e9cf308-c94b-4a76-9258-93c5d7563146", "address": "fa:16:3e:d1:6b:5e", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e9cf308-c9", "ovs_interfaceid": "2e9cf308-c94b-4a76-9258-93c5d7563146", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 875.419328] env[65385]: DEBUG nova.compute.utils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 875.420916] env[65385]: DEBUG nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 875.421129] env[65385]: DEBUG nova.network.neutron [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 875.421594] env[65385]: WARNING neutronclient.v2_0.client [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 875.422292] env[65385]: WARNING neutronclient.v2_0.client [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 875.423434] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.423814] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.488253] env[65385]: DEBUG nova.policy [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '04b763d9de2d4755a85e2a3b4d09d677', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8519227171034e9fa29a28cd5f94f6a3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 875.787815] env[65385]: DEBUG nova.network.neutron [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Successfully created port: 15a6f057-527a-433e-bb76-8d2b57f48765 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 875.934067] env[65385]: DEBUG nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 876.259273] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 876.259556] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.259709] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 876.259990] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.260350] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 876.260350] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 876.260512] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.260652] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 876.260749] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 876.260910] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 876.261093] env[65385]: DEBUG nova.virt.hardware [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 876.263708] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300e5b66-8c50-4c69-9e9a-ba05dac26f2c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.277031] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52385975-7571-e4eb-5b74-9c3277a0c606/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 876.278907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f45c37-fd32-42fb-9f7a-479e47f26603 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.292026] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6643f57f-f560-4330-81e7-607348a86d2c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.296777] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52385975-7571-e4eb-5b74-9c3277a0c606/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 876.297062] env[65385]: ERROR oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52385975-7571-e4eb-5b74-9c3277a0c606/disk-0.vmdk due to incomplete transfer. [ 876.297869] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a7b57e73-4467-4508-a788-c879be10987e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.301533] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c12d18-c6fd-47dd-968e-ae1dd291a5d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.313517] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:6b:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b89fd3b-0470-40c9-bb5b-d52c76c030e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e9cf308-c94b-4a76-9258-93c5d7563146', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.321876] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 876.323667] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.323949] env[65385]: DEBUG oslo_vmware.rw_handles [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52385975-7571-e4eb-5b74-9c3277a0c606/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 876.324132] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Uploaded image c3f38eb9-8503-49ea-a6f5-d7759c1ac49c to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 876.326764] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 876.329141] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f151268-444a-460b-a4cd-cc71a18a6be4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.343818] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-216946b5-00ba-4a10-bde4-5a6c4815ce7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.346605] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c345e3-837c-413b-9346-90f08759a832 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.385190] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0496c631-47f3-4a86-a757-d64674d90f71 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.388089] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.388089] env[65385]: value = "task-4453874" [ 876.388089] env[65385]: _type = "Task" [ 876.388089] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.388353] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 876.388353] env[65385]: value = "task-4453875" [ 876.388353] env[65385]: _type = "Task" [ 876.388353] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.399498] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f91e5d7-8094-4702-ba53-e13425874d90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.410621] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453875, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.410902] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453874, 'name': CreateVM_Task} progress is 15%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.422825] env[65385]: DEBUG nova.compute.provider_tree [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.603056] env[65385]: DEBUG nova.compute.manager [req-ce52569e-2432-4805-8402-182fe797e332 req-c774ee5a-e8c1-4ce0-aed7-a81286d71cdf service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Received event network-vif-plugged-2e9cf308-c94b-4a76-9258-93c5d7563146 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 876.603449] env[65385]: DEBUG oslo_concurrency.lockutils [req-ce52569e-2432-4805-8402-182fe797e332 req-c774ee5a-e8c1-4ce0-aed7-a81286d71cdf service nova] Acquiring lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 876.604633] env[65385]: DEBUG oslo_concurrency.lockutils [req-ce52569e-2432-4805-8402-182fe797e332 req-c774ee5a-e8c1-4ce0-aed7-a81286d71cdf service nova] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.604633] env[65385]: DEBUG oslo_concurrency.lockutils [req-ce52569e-2432-4805-8402-182fe797e332 req-c774ee5a-e8c1-4ce0-aed7-a81286d71cdf service nova] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.604633] env[65385]: DEBUG nova.compute.manager [req-ce52569e-2432-4805-8402-182fe797e332 req-c774ee5a-e8c1-4ce0-aed7-a81286d71cdf service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] No waiting events found dispatching network-vif-plugged-2e9cf308-c94b-4a76-9258-93c5d7563146 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 876.604633] env[65385]: WARNING nova.compute.manager [req-ce52569e-2432-4805-8402-182fe797e332 req-c774ee5a-e8c1-4ce0-aed7-a81286d71cdf service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Received unexpected event network-vif-plugged-2e9cf308-c94b-4a76-9258-93c5d7563146 for instance with vm_state building and task_state spawning. [ 876.901910] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453874, 'name': CreateVM_Task, 'duration_secs': 0.526172} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.905065] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.905503] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453875, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.906095] env[65385]: WARNING neutronclient.v2_0.client [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.906558] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.906978] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.907421] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 876.907722] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49ef1752-6fe4-4360-9142-10b19b1e905d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.914095] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 876.914095] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528f0683-d240-2b05-8d88-64715c01b130" [ 876.914095] env[65385]: _type = "Task" [ 876.914095] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.924377] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528f0683-d240-2b05-8d88-64715c01b130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.926516] env[65385]: DEBUG nova.scheduler.client.report [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 876.945986] env[65385]: DEBUG nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 876.983551] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 876.983922] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.984218] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 876.984454] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.984622] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 876.984802] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 876.985064] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.985237] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 876.985406] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 876.985566] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 876.985736] env[65385]: DEBUG nova.virt.hardware [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 876.987236] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30d219c-a5d0-4495-bd70-aa3996f672ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.998452] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc62db26-61ad-4152-b3ad-4ef678ee4c1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.106525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 877.106999] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.107374] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 877.107633] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.107851] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.110058] env[65385]: INFO nova.compute.manager [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Terminating instance [ 877.249990] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5292678e-0fac-a45b-ab8e-472e49b928a6/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 877.251650] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89706fc0-bcb0-47e6-977e-969ed8b8d670 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.258946] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5292678e-0fac-a45b-ab8e-472e49b928a6/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 877.259242] env[65385]: ERROR oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5292678e-0fac-a45b-ab8e-472e49b928a6/disk-0.vmdk due to incomplete transfer. [ 877.260036] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6ddd918a-b873-4dbb-bfbd-ea4dab712d2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.269164] env[65385]: DEBUG oslo_vmware.rw_handles [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5292678e-0fac-a45b-ab8e-472e49b928a6/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 877.269932] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Uploaded image 57fab5fe-ac7e-4d91-94ca-0285c2957e33 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 877.272736] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 877.273039] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6bf8e38e-c74c-4749-935e-e0fe940b4731 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.281373] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 877.281373] env[65385]: value = "task-4453876" [ 877.281373] env[65385]: _type = "Task" [ 877.281373] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.292717] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453876, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.400745] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453875, 'name': Destroy_Task, 'duration_secs': 0.630529} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.401031] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Destroyed the VM [ 877.401531] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 877.401869] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c84591b7-1219-4058-8465-1e709aa0bb05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.412014] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 877.412014] env[65385]: value = "task-4453877" [ 877.412014] env[65385]: _type = "Task" [ 877.412014] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.426068] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453877, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.434147] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528f0683-d240-2b05-8d88-64715c01b130, 'name': SearchDatastore_Task, 'duration_secs': 0.017663} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.434147] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.434147] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.434147] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.434147] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.434147] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.434464] env[65385]: DEBUG nova.network.neutron [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Successfully updated port: 15a6f057-527a-433e-bb76-8d2b57f48765 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 877.438976] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbf7baeb-ea06-4a40-a1d0-e0df0078bfc7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.444615] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.444615] env[65385]: DEBUG nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 877.449529] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.757s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.450097] env[65385]: DEBUG nova.objects.instance [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'resources' on Instance uuid b1cca536-e1d9-43f8-8a5c-be3981c106e7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.460931] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.461346] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.465450] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e7bdc8c-7265-4688-a3de-4541799fa3e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.477233] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 877.477233] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5256fa4f-7ad3-0512-d0f7-9b68a90269e8" [ 877.477233] env[65385]: _type = "Task" [ 877.477233] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.489521] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5256fa4f-7ad3-0512-d0f7-9b68a90269e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.614463] env[65385]: DEBUG nova.compute.manager [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 877.614463] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.615691] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927d68d8-0a74-4f7f-bd76-0a1d30897710 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.625261] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.625610] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5645b3f-1fce-4eee-b7d5-b7e44dfe1a2c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.634633] env[65385]: DEBUG oslo_vmware.api [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 877.634633] env[65385]: value = "task-4453878" [ 877.634633] env[65385]: _type = "Task" [ 877.634633] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.645681] env[65385]: DEBUG oslo_vmware.api [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.744933] env[65385]: DEBUG nova.compute.manager [req-fd49a43e-ff22-4d0c-b67e-ee8abaf7d0e4 req-e6a82003-f683-4ca1-9319-bc3b75b8d476 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Received event network-vif-plugged-15a6f057-527a-433e-bb76-8d2b57f48765 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 877.745230] env[65385]: DEBUG oslo_concurrency.lockutils [req-fd49a43e-ff22-4d0c-b67e-ee8abaf7d0e4 req-e6a82003-f683-4ca1-9319-bc3b75b8d476 service nova] Acquiring lock "6dae15ad-a340-47e3-ab6b-95b25043ca27-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 877.745441] env[65385]: DEBUG oslo_concurrency.lockutils [req-fd49a43e-ff22-4d0c-b67e-ee8abaf7d0e4 req-e6a82003-f683-4ca1-9319-bc3b75b8d476 service nova] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.745687] env[65385]: DEBUG oslo_concurrency.lockutils [req-fd49a43e-ff22-4d0c-b67e-ee8abaf7d0e4 req-e6a82003-f683-4ca1-9319-bc3b75b8d476 service nova] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.745777] env[65385]: DEBUG nova.compute.manager [req-fd49a43e-ff22-4d0c-b67e-ee8abaf7d0e4 req-e6a82003-f683-4ca1-9319-bc3b75b8d476 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] No waiting events found dispatching network-vif-plugged-15a6f057-527a-433e-bb76-8d2b57f48765 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 877.745936] env[65385]: WARNING nova.compute.manager [req-fd49a43e-ff22-4d0c-b67e-ee8abaf7d0e4 req-e6a82003-f683-4ca1-9319-bc3b75b8d476 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Received unexpected event network-vif-plugged-15a6f057-527a-433e-bb76-8d2b57f48765 for instance with vm_state building and task_state spawning. [ 877.794887] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453876, 'name': Destroy_Task, 'duration_secs': 0.351445} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.795216] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Destroyed the VM [ 877.795480] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 877.795821] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-30bdfcf2-b1c3-4ad8-9fc8-caf4f5d22dc0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.804809] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 877.804809] env[65385]: value = "task-4453879" [ 877.804809] env[65385]: _type = "Task" [ 877.804809] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.816267] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453879, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.924091] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453877, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.950151] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "refresh_cache-6dae15ad-a340-47e3-ab6b-95b25043ca27" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.950481] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquired lock "refresh_cache-6dae15ad-a340-47e3-ab6b-95b25043ca27" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.950998] env[65385]: DEBUG nova.network.neutron [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 877.955441] env[65385]: DEBUG nova.compute.utils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 877.957415] env[65385]: DEBUG nova.objects.instance [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'numa_topology' on Instance uuid b1cca536-e1d9-43f8-8a5c-be3981c106e7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.958800] env[65385]: DEBUG nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 877.959065] env[65385]: DEBUG nova.network.neutron [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 877.959474] env[65385]: WARNING neutronclient.v2_0.client [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.959842] env[65385]: WARNING neutronclient.v2_0.client [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.960491] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.960886] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.969740] env[65385]: DEBUG nova.objects.base [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 877.989166] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5256fa4f-7ad3-0512-d0f7-9b68a90269e8, 'name': SearchDatastore_Task, 'duration_secs': 0.017869} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.990435] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45c5afc9-0065-4bcf-8097-e69637e308d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.997324] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 877.997324] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0b56f-fc18-2c9d-5eb3-80f8124d35e7" [ 877.997324] env[65385]: _type = "Task" [ 877.997324] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.007326] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0b56f-fc18-2c9d-5eb3-80f8124d35e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.021218] env[65385]: DEBUG nova.policy [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '400a27995d0e41df9cb945625fa4e461', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21234ef9798c48278f3f65cdd9913f3c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 878.149620] env[65385]: DEBUG oslo_vmware.api [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453878, 'name': PowerOffVM_Task, 'duration_secs': 0.26649} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.149915] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.150102] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.150393] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e475205a-eed7-4a19-8dbf-818e9b4b4e30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.226773] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 878.229163] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 878.229474] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Deleting the datastore file [datastore2] 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.229809] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a879945b-eed8-4aa6-9413-161660b0b856 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.240034] env[65385]: DEBUG oslo_vmware.api [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for the task: (returnval){ [ 878.240034] env[65385]: value = "task-4453881" [ 878.240034] env[65385]: _type = "Task" [ 878.240034] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.252774] env[65385]: DEBUG oslo_vmware.api [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453881, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.323521] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453879, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.395226] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223d92b1-13c6-42c7-91bc-0c691e0e48af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.404435] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520a0115-e412-4cf9-83c7-ca4bb1e09134 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.445782] env[65385]: DEBUG nova.network.neutron [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Successfully created port: ad2419b0-bd5b-4675-b73e-a9afc321bf4c {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 878.451964] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42a356b-003a-4f6e-bc91-a297228e5a42 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.462701] env[65385]: DEBUG nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 878.465240] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.465606] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.473671] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453877, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.482805] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8518101-550c-4af2-b10c-3c6446c167ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.504666] env[65385]: DEBUG nova.compute.provider_tree [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.518227] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0b56f-fc18-2c9d-5eb3-80f8124d35e7, 'name': SearchDatastore_Task, 'duration_secs': 0.017632} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.518428] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.518930] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] e02e3848-1b1e-426b-bc0f-24c8e232a9db/e02e3848-1b1e-426b-bc0f-24c8e232a9db.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.519098] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1d92534-b9c7-4edc-be51-77966ef0df63 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.529194] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 878.529194] env[65385]: value = "task-4453882" [ 878.529194] env[65385]: _type = "Task" [ 878.529194] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.541324] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.680848] env[65385]: DEBUG nova.network.neutron [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 878.752365] env[65385]: DEBUG oslo_vmware.api [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Task: {'id': task-4453881, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.422476} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.752670] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.752893] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.753172] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.753389] env[65385]: INFO nova.compute.manager [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 878.753646] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 878.753851] env[65385]: DEBUG nova.compute.manager [-] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 878.754029] env[65385]: DEBUG nova.network.neutron [-] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 878.754340] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.755949] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.755949] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.821091] env[65385]: DEBUG oslo_vmware.api [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453879, 'name': RemoveSnapshot_Task, 'duration_secs': 0.76325} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.821453] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 878.821704] env[65385]: INFO nova.compute.manager [None req-500d9d02-1e29-4498-b3bf-9336ce4424c8 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Took 12.96 seconds to snapshot the instance on the hypervisor. [ 878.872123] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.872599] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.928248] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.955483] env[65385]: DEBUG oslo_vmware.api [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453877, 'name': RemoveSnapshot_Task, 'duration_secs': 1.061617} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.955775] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 878.956024] env[65385]: INFO nova.compute.manager [None req-d8a64508-29e1-406e-9dd6-448227e88a0f tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Took 15.81 seconds to snapshot the instance on the hypervisor. [ 878.989169] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "0b02fcd5-9fc7-4543-a754-d31fa0186981" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.989500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.989764] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "0b02fcd5-9fc7-4543-a754-d31fa0186981-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.989967] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.990152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.992433] env[65385]: INFO nova.compute.manager [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Terminating instance [ 879.010636] env[65385]: DEBUG nova.scheduler.client.report [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 879.042952] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453882, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.482775] env[65385]: DEBUG nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 879.500277] env[65385]: DEBUG nova.compute.manager [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 879.500612] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 879.502247] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad41bfc-0784-4591-b395-e07ea3d7906d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.515343] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.521042] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d989d04d-8b43-4320-ba5d-bfcd58ea763c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.522066] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.072s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.530174] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.048s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.530521] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.532731] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.030s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.534240] env[65385]: INFO nova.compute.claims [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.536905] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 879.537137] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.540720] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 879.540955] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 879.541122] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 879.541299] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 879.541454] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 879.541583] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 879.541817] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.541929] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 879.542252] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 879.542252] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 879.542592] env[65385]: DEBUG nova.virt.hardware [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 879.544530] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-003f8a9c-84d6-42fc-b0b6-bf53c8f6cf9b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.550479] env[65385]: DEBUG nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 879.560546] env[65385]: DEBUG oslo_vmware.api [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 879.560546] env[65385]: value = "task-4453883" [ 879.560546] env[65385]: _type = "Task" [ 879.560546] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.576612] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453882, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.868272} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.577929] env[65385]: INFO nova.scheduler.client.report [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleted allocations for instance 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22 [ 879.582644] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] e02e3848-1b1e-426b-bc0f-24c8e232a9db/e02e3848-1b1e-426b-bc0f-24c8e232a9db.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.582644] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.582644] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac48d6d1-a4f5-47be-8c53-d2830746d2f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.594924] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b496e14b-411e-4313-b723-8fe78987dd61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.597703] env[65385]: DEBUG oslo_vmware.api [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453883, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.617224] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 879.617224] env[65385]: value = "task-4453884" [ 879.617224] env[65385]: _type = "Task" [ 879.617224] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.854283] env[65385]: WARNING neutronclient.v2_0.client [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 879.855129] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.855173] env[65385]: WARNING openstack [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.033443] env[65385]: DEBUG oslo_concurrency.lockutils [None req-58dbc89c-4a81-469f-afc5-1a01540c5ebc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 43.551s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.034361] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 18.382s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.034588] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.034785] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.034942] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.040658] env[65385]: INFO nova.compute.manager [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Terminating instance [ 880.077957] env[65385]: DEBUG oslo_vmware.api [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453883, 'name': PowerOffVM_Task, 'duration_secs': 0.297502} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.078267] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.078432] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.078711] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-489b246b-48da-4104-9ee8-06a4832e969b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.086299] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.103662] env[65385]: DEBUG oslo_concurrency.lockutils [None req-81e44b55-5a17-4d5c-9cc7-c1d18d22a114 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "38e0ebf5-1d92-4906-b0bd-b0ebf8651f22" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.299s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.126452] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453884, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093128} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.127127] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.127627] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d792f355-18f2-44a0-a9ba-8b5b8af5b055 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.155316] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] e02e3848-1b1e-426b-bc0f-24c8e232a9db/e02e3848-1b1e-426b-bc0f-24c8e232a9db.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.156910] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7ed08d1-889e-4a8e-8a90-dd2b6723c875 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.172303] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 880.172558] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 880.172684] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Deleting the datastore file [datastore1] 0b02fcd5-9fc7-4543-a754-d31fa0186981 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.173387] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad9357c8-3f54-4f32-8e6c-6bc39d836aa3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.182891] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 880.182891] env[65385]: value = "task-4453887" [ 880.182891] env[65385]: _type = "Task" [ 880.182891] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.183196] env[65385]: DEBUG oslo_vmware.api [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for the task: (returnval){ [ 880.183196] env[65385]: value = "task-4453886" [ 880.183196] env[65385]: _type = "Task" [ 880.183196] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.195960] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453887, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.199270] env[65385]: DEBUG oslo_vmware.api [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453886, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.219440] env[65385]: DEBUG nova.network.neutron [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Successfully updated port: ad2419b0-bd5b-4675-b73e-a9afc321bf4c {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 880.225781] env[65385]: DEBUG nova.network.neutron [-] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 880.334744] env[65385]: DEBUG nova.network.neutron [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Updating instance_info_cache with network_info: [{"id": "15a6f057-527a-433e-bb76-8d2b57f48765", "address": "fa:16:3e:da:46:5d", "network": {"id": "84c9fea3-777d-4976-bc50-5de50f5221f9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1134281187-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8519227171034e9fa29a28cd5f94f6a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22390021-1742-415d-b442-811550d09927", "external-id": "nsx-vlan-transportzone-347", "segmentation_id": 347, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15a6f057-52", "ovs_interfaceid": "15a6f057-527a-433e-bb76-8d2b57f48765", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 880.409970] env[65385]: DEBUG nova.compute.manager [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Received event network-changed-2e9cf308-c94b-4a76-9258-93c5d7563146 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 880.410254] env[65385]: DEBUG nova.compute.manager [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Refreshing instance network info cache due to event network-changed-2e9cf308-c94b-4a76-9258-93c5d7563146. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 880.410578] env[65385]: DEBUG oslo_concurrency.lockutils [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Acquiring lock "refresh_cache-e02e3848-1b1e-426b-bc0f-24c8e232a9db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.410812] env[65385]: DEBUG oslo_concurrency.lockutils [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Acquired lock "refresh_cache-e02e3848-1b1e-426b-bc0f-24c8e232a9db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.411087] env[65385]: DEBUG nova.network.neutron [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Refreshing network info cache for port 2e9cf308-c94b-4a76-9258-93c5d7563146 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 880.546350] env[65385]: DEBUG nova.compute.manager [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 880.546702] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.551025] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db641ec6-93fb-461a-9d4a-27247fde5afd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.563804] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcf4f23-2d65-4070-8f9b-cd30183f7ad1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.602959] env[65385]: WARNING nova.virt.vmwareapi.vmops [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1cca536-e1d9-43f8-8a5c-be3981c106e7 could not be found. [ 880.603194] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.603383] env[65385]: INFO nova.compute.manager [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Took 0.06 seconds to destroy the instance on the hypervisor. [ 880.603635] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 880.606970] env[65385]: DEBUG nova.compute.manager [-] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 880.607095] env[65385]: DEBUG nova.network.neutron [-] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 880.607339] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.607860] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.608134] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.696881] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.705542] env[65385]: DEBUG oslo_vmware.api [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Task: {'id': task-4453886, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257127} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.705542] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453887, 'name': ReconfigVM_Task, 'duration_secs': 0.397052} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.708162] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.708362] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.708534] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.708709] env[65385]: INFO nova.compute.manager [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Took 1.21 seconds to destroy the instance on the hypervisor. [ 880.708944] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 880.709168] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Reconfigured VM instance instance-00000044 to attach disk [datastore1] e02e3848-1b1e-426b-bc0f-24c8e232a9db/e02e3848-1b1e-426b-bc0f-24c8e232a9db.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.710076] env[65385]: DEBUG nova.compute.manager [-] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 880.710327] env[65385]: DEBUG nova.network.neutron [-] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 880.710410] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.711192] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.711342] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.717396] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a40a56b7-6c3e-4d9f-adc3-5852c656f028 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.722392] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "refresh_cache-fa9a5278-1477-485a-9201-a37187488aef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.722560] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquired lock "refresh_cache-fa9a5278-1477-485a-9201-a37187488aef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.722817] env[65385]: DEBUG nova.network.neutron [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 880.727393] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 880.727393] env[65385]: value = "task-4453888" [ 880.727393] env[65385]: _type = "Task" [ 880.727393] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.731074] env[65385]: INFO nova.compute.manager [-] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Took 1.98 seconds to deallocate network for instance. [ 880.746464] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453888, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.839278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Releasing lock "refresh_cache-6dae15ad-a340-47e3-ab6b-95b25043ca27" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 880.839726] env[65385]: DEBUG nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Instance network_info: |[{"id": "15a6f057-527a-433e-bb76-8d2b57f48765", "address": "fa:16:3e:da:46:5d", "network": {"id": "84c9fea3-777d-4976-bc50-5de50f5221f9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1134281187-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8519227171034e9fa29a28cd5f94f6a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22390021-1742-415d-b442-811550d09927", "external-id": "nsx-vlan-transportzone-347", "segmentation_id": 347, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15a6f057-52", "ovs_interfaceid": "15a6f057-527a-433e-bb76-8d2b57f48765", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 880.840230] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:46:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '22390021-1742-415d-b442-811550d09927', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15a6f057-527a-433e-bb76-8d2b57f48765', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.848185] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Creating folder: Project (8519227171034e9fa29a28cd5f94f6a3). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.851583] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d1ac245-52e0-4a21-a463-c61e3cd9fb8e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.865089] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Created folder: Project (8519227171034e9fa29a28cd5f94f6a3) in parent group-v870881. [ 880.865355] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Creating folder: Instances. Parent ref: group-v871080. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.868478] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61340695-05b3-4d1a-88ee-997fc1db3fa2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.880846] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Created folder: Instances in parent group-v871080. [ 880.881139] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 880.881336] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.881570] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d8fab5c-2f51-43da-95b4-53f075925d48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.905988] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.905988] env[65385]: value = "task-4453891" [ 880.905988] env[65385]: _type = "Task" [ 880.905988] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.914040] env[65385]: WARNING neutronclient.v2_0.client [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.915099] env[65385]: WARNING openstack [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.915454] env[65385]: WARNING openstack [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.928644] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453891, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.005534] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.050852] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e22c949-a37e-4684-a437-d755505e1ca6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.058744] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "ac6fa516-d954-466e-b8e0-e12440492049" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.059040] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "ac6fa516-d954-466e-b8e0-e12440492049" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 881.064152] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45c8ae7-a86f-4573-9101-7d6fc77cc1f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.114922] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41db1c44-57d0-423b-8f32-f20eddb30ee4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.125219] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b0bb4c-36ab-4696-bf8d-144a846431a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.143221] env[65385]: DEBUG nova.compute.provider_tree [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.225862] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.226260] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.238535] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.246986] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453888, 'name': Rename_Task, 'duration_secs': 0.149477} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.247291] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.247584] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b2af199-9b90-4980-b105-e6893b375d3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.255531] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 881.255531] env[65385]: value = "task-4453892" [ 881.255531] env[65385]: _type = "Task" [ 881.255531] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.273170] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.360343] env[65385]: DEBUG nova.network.neutron [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 881.379703] env[65385]: WARNING openstack [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.381033] env[65385]: WARNING openstack [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.414837] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.415542] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.435790] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453891, 'name': CreateVM_Task, 'duration_secs': 0.370883} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.435790] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.435790] env[65385]: WARNING neutronclient.v2_0.client [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.435790] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.437747] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.437747] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 881.437747] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d46a282-c33c-42c0-a76c-1726643c12c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.443864] env[65385]: DEBUG nova.network.neutron [-] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.445354] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 881.445354] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522ccdd9-4b51-2080-bcc3-f2615ef42f6b" [ 881.445354] env[65385]: _type = "Task" [ 881.445354] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.457611] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522ccdd9-4b51-2080-bcc3-f2615ef42f6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.546969] env[65385]: WARNING neutronclient.v2_0.client [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.547742] env[65385]: WARNING openstack [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.548128] env[65385]: WARNING openstack [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.563862] env[65385]: DEBUG nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 881.611133] env[65385]: WARNING neutronclient.v2_0.client [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.611133] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.611133] env[65385]: WARNING openstack [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.666763] env[65385]: ERROR nova.scheduler.client.report [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [req-008c5c77-93fe-4041-bce3-ace4f2ce05cc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-008c5c77-93fe-4041-bce3-ace4f2ce05cc"}]} [ 881.688404] env[65385]: DEBUG nova.scheduler.client.report [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 881.706871] env[65385]: DEBUG nova.scheduler.client.report [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 881.707115] env[65385]: DEBUG nova.compute.provider_tree [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.721053] env[65385]: DEBUG nova.scheduler.client.report [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 881.743171] env[65385]: DEBUG nova.scheduler.client.report [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 881.770225] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453892, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.809082] env[65385]: DEBUG nova.network.neutron [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Updated VIF entry in instance network info cache for port 2e9cf308-c94b-4a76-9258-93c5d7563146. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 881.810143] env[65385]: DEBUG nova.network.neutron [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Updating instance_info_cache with network_info: [{"id": "2e9cf308-c94b-4a76-9258-93c5d7563146", "address": "fa:16:3e:d1:6b:5e", "network": {"id": "55c148e0-fc9c-4298-90a5-65d172e909f6", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-387749867-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99f640b7d5594400bbd69c05713089d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e9cf308-c9", "ovs_interfaceid": "2e9cf308-c94b-4a76-9258-93c5d7563146", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.834416] env[65385]: DEBUG nova.network.neutron [-] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.841171] env[65385]: DEBUG nova.network.neutron [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Updating instance_info_cache with network_info: [{"id": "ad2419b0-bd5b-4675-b73e-a9afc321bf4c", "address": "fa:16:3e:f7:99:c9", "network": {"id": "5da125f0-44fc-462c-8a94-96a6f22a3a68", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-467202416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21234ef9798c48278f3f65cdd9913f3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad2419b0-bd", "ovs_interfaceid": "ad2419b0-bd5b-4675-b73e-a9afc321bf4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.908659] env[65385]: DEBUG nova.compute.manager [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Received event network-changed-15a6f057-527a-433e-bb76-8d2b57f48765 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 881.908857] env[65385]: DEBUG nova.compute.manager [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Refreshing instance network info cache due to event network-changed-15a6f057-527a-433e-bb76-8d2b57f48765. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 881.909079] env[65385]: DEBUG oslo_concurrency.lockutils [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Acquiring lock "refresh_cache-6dae15ad-a340-47e3-ab6b-95b25043ca27" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.909252] env[65385]: DEBUG oslo_concurrency.lockutils [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Acquired lock "refresh_cache-6dae15ad-a340-47e3-ab6b-95b25043ca27" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.909369] env[65385]: DEBUG nova.network.neutron [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Refreshing network info cache for port 15a6f057-527a-433e-bb76-8d2b57f48765 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 881.950917] env[65385]: INFO nova.compute.manager [-] [instance: b1cca536-e1d9-43f8-8a5c-be3981c106e7] Took 1.34 seconds to deallocate network for instance. [ 881.960576] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522ccdd9-4b51-2080-bcc3-f2615ef42f6b, 'name': SearchDatastore_Task, 'duration_secs': 0.012364} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.963763] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.964140] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.964373] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.964608] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.964731] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.965291] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc6b365b-2e15-49b7-bb14-96cc6ea49c8d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.983394] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.983537] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.984399] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d88738c-349c-4d0e-b212-a9b5e64c7ea7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.991883] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 881.991883] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5235ae12-c593-f7f9-6723-6bb99d5e1a40" [ 881.991883] env[65385]: _type = "Task" [ 881.991883] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.007422] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5235ae12-c593-f7f9-6723-6bb99d5e1a40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.088229] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.226290] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2593889f-86e2-48b1-9015-8cda004da2b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.235230] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1479695d-bb3b-4ec1-a797-71a80f278433 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.274321] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a180c623-1ec1-419c-8917-9d81e2239fba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.292061] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b19aee4-dcd2-4910-b823-bfcb47525adb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.298361] env[65385]: DEBUG oslo_vmware.api [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453892, 'name': PowerOnVM_Task, 'duration_secs': 0.546469} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.298361] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.298477] env[65385]: INFO nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Took 10.06 seconds to spawn the instance on the hypervisor. [ 882.298545] env[65385]: DEBUG nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 882.300421] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2659809a-1dc7-4d64-bfb4-3bce5187715c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.313572] env[65385]: DEBUG oslo_concurrency.lockutils [req-032f2b18-c092-49bf-a2f6-508a8bf85df5 req-c761fe22-a48d-47ae-835f-3948036fc5e8 service nova] Releasing lock "refresh_cache-e02e3848-1b1e-426b-bc0f-24c8e232a9db" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.314314] env[65385]: DEBUG nova.compute.provider_tree [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.337083] env[65385]: INFO nova.compute.manager [-] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Took 1.63 seconds to deallocate network for instance. [ 882.349981] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Releasing lock "refresh_cache-fa9a5278-1477-485a-9201-a37187488aef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.349981] env[65385]: DEBUG nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Instance network_info: |[{"id": "ad2419b0-bd5b-4675-b73e-a9afc321bf4c", "address": "fa:16:3e:f7:99:c9", "network": {"id": "5da125f0-44fc-462c-8a94-96a6f22a3a68", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-467202416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21234ef9798c48278f3f65cdd9913f3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad2419b0-bd", "ovs_interfaceid": "ad2419b0-bd5b-4675-b73e-a9afc321bf4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 882.349981] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:99:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cf5bfbae-a882-4d34-be33-b31e274b3077', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad2419b0-bd5b-4675-b73e-a9afc321bf4c', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.361073] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Creating folder: Project (21234ef9798c48278f3f65cdd9913f3c). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.361426] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4534079-1a5a-47d5-8705-26333fcbbbbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.375106] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Created folder: Project (21234ef9798c48278f3f65cdd9913f3c) in parent group-v870881. [ 882.375327] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Creating folder: Instances. Parent ref: group-v871083. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.375569] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acb8605d-b68f-419f-a78b-aaae31c10715 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.389615] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Created folder: Instances in parent group-v871083. [ 882.389896] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 882.390177] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa9a5278-1477-485a-9201-a37187488aef] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.390429] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bac73ba8-adcb-42dd-bdca-13b270dfff58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.413305] env[65385]: WARNING neutronclient.v2_0.client [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.414313] env[65385]: WARNING openstack [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.415144] env[65385]: WARNING openstack [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.428956] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.428956] env[65385]: value = "task-4453895" [ 882.428956] env[65385]: _type = "Task" [ 882.428956] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.444310] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453895, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.505842] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5235ae12-c593-f7f9-6723-6bb99d5e1a40, 'name': SearchDatastore_Task, 'duration_secs': 0.032779} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.507826] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee2885bc-3893-4936-bbfb-a5d8be1c52e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.517339] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 882.517339] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52557a7a-cabb-126b-2d2d-ad572e5cae5e" [ 882.517339] env[65385]: _type = "Task" [ 882.517339] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.527474] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52557a7a-cabb-126b-2d2d-ad572e5cae5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.671473] env[65385]: WARNING openstack [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.672645] env[65385]: WARNING openstack [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.743691] env[65385]: WARNING neutronclient.v2_0.client [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.744415] env[65385]: WARNING openstack [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.744810] env[65385]: WARNING openstack [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.756314] env[65385]: DEBUG nova.compute.manager [req-32238e62-db05-42f4-a69f-bf02a09b9590 req-fb51b648-ae38-4057-ba86-251660619fb4 service nova] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Received event network-vif-deleted-0ee1cb97-07c8-4c45-9442-d76deacb6028 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 882.838599] env[65385]: INFO nova.compute.manager [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Took 36.51 seconds to build instance. [ 882.843566] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.866321] env[65385]: DEBUG nova.network.neutron [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Updated VIF entry in instance network info cache for port 15a6f057-527a-433e-bb76-8d2b57f48765. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 882.866864] env[65385]: DEBUG nova.network.neutron [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Updating instance_info_cache with network_info: [{"id": "15a6f057-527a-433e-bb76-8d2b57f48765", "address": "fa:16:3e:da:46:5d", "network": {"id": "84c9fea3-777d-4976-bc50-5de50f5221f9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1134281187-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8519227171034e9fa29a28cd5f94f6a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22390021-1742-415d-b442-811550d09927", "external-id": "nsx-vlan-transportzone-347", "segmentation_id": 347, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15a6f057-52", "ovs_interfaceid": "15a6f057-527a-433e-bb76-8d2b57f48765", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 882.888558] env[65385]: DEBUG nova.scheduler.client.report [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Updated inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with generation 96 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 882.888806] env[65385]: DEBUG nova.compute.provider_tree [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Updating resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b generation from 96 to 97 during operation: update_inventory {{(pid=65385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 882.889156] env[65385]: DEBUG nova.compute.provider_tree [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.943334] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453895, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.999911] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e53aa0cf-5f77-43ed-9139-e9103542d911 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "b1cca536-e1d9-43f8-8a5c-be3981c106e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.965s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.027802] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52557a7a-cabb-126b-2d2d-ad572e5cae5e, 'name': SearchDatastore_Task, 'duration_secs': 0.01841} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.030201] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.030476] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 6dae15ad-a340-47e3-ab6b-95b25043ca27/6dae15ad-a340-47e3-ab6b-95b25043ca27.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.033611] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bad23da0-59aa-4015-b6c3-b9aec38d54f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.047038] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 883.047038] env[65385]: value = "task-4453896" [ 883.047038] env[65385]: _type = "Task" [ 883.047038] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.058081] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.340693] env[65385]: DEBUG oslo_concurrency.lockutils [None req-783a0a1f-7a87-4047-9248-1b0c01e1286e tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.016s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.372625] env[65385]: DEBUG oslo_concurrency.lockutils [req-bc1725a9-b889-47dd-afa1-5b04cdc9fdde req-e298d0e9-af72-4fa2-bad0-b69ecdd8f123 service nova] Releasing lock "refresh_cache-6dae15ad-a340-47e3-ab6b-95b25043ca27" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.396932] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.864s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.397408] env[65385]: DEBUG nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 883.400208] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.507s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.400418] env[65385]: DEBUG nova.objects.instance [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lazy-loading 'resources' on Instance uuid 7fcbdc79-688c-479f-94e2-f4542abe714b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.446726] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453895, 'name': CreateVM_Task, 'duration_secs': 0.60492} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.446726] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa9a5278-1477-485a-9201-a37187488aef] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 883.446726] env[65385]: WARNING neutronclient.v2_0.client [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.446726] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.446726] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.446726] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 883.449026] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05f89f14-055a-4656-a61f-1a278f19e4ae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.453650] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 883.453650] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290134e-3402-70fb-028e-c5cb60197d5e" [ 883.453650] env[65385]: _type = "Task" [ 883.453650] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.462696] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290134e-3402-70fb-028e-c5cb60197d5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.471687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "42b9edf6-1873-49c3-8074-8eef654ac371" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 883.471687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.471687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 883.471687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.471687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.474243] env[65385]: INFO nova.compute.manager [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Terminating instance [ 883.537659] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 883.537957] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.538210] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 883.538410] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.538603] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.547402] env[65385]: INFO nova.compute.manager [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Terminating instance [ 883.564982] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453896, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.903301] env[65385]: DEBUG nova.compute.utils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 883.907660] env[65385]: DEBUG nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 883.907910] env[65385]: DEBUG nova.network.neutron [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 883.908199] env[65385]: WARNING neutronclient.v2_0.client [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.908491] env[65385]: WARNING neutronclient.v2_0.client [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.909083] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.909436] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.965153] env[65385]: DEBUG nova.policy [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64999b5f66fe4e06ae1ad0d9822f0352', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1187e438e194c4587f8824110dc23f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 883.974271] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290134e-3402-70fb-028e-c5cb60197d5e, 'name': SearchDatastore_Task, 'duration_secs': 0.011795} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.977874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.978209] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.978488] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.978702] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.978968] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.979855] env[65385]: DEBUG nova.compute.manager [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 883.980246] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.980521] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-735cd5c6-4c81-4a8e-9e93-9b205d3f4f62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.982986] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c194eb-29be-4f8b-8e70-31e119f66553 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.993016] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.993016] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6f663bc-7dc3-4396-b57a-46bf590d7566 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.996018] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.996148] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.999809] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7576697-3a99-494e-91de-fd2ac4c1ec9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.008908] env[65385]: DEBUG oslo_vmware.api [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 884.008908] env[65385]: value = "task-4453897" [ 884.008908] env[65385]: _type = "Task" [ 884.008908] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.009830] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 884.009830] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52acd029-c95b-01d3-b114-d536e1d07547" [ 884.009830] env[65385]: _type = "Task" [ 884.009830] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.028054] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52acd029-c95b-01d3-b114-d536e1d07547, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.028054] env[65385]: DEBUG oslo_vmware.api [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453897, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.059612] env[65385]: DEBUG nova.compute.manager [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 884.059681] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 884.060374] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.060631] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.068552] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c713203-e49f-4b52-8ba7-be3ad3f1cccd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.072337] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453896, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.827571} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.072948] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 6dae15ad-a340-47e3-ab6b-95b25043ca27/6dae15ad-a340-47e3-ab6b-95b25043ca27.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 884.073182] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.073826] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3710a941-b4ca-4d85-b3e2-ac8c0bd21cbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.080065] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.080380] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9926d73-de59-401f-acf3-93257eb0cfd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.083910] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 884.083910] env[65385]: value = "task-4453898" [ 884.083910] env[65385]: _type = "Task" [ 884.083910] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.095270] env[65385]: DEBUG oslo_vmware.api [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 884.095270] env[65385]: value = "task-4453899" [ 884.095270] env[65385]: _type = "Task" [ 884.095270] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.103595] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453898, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.114417] env[65385]: DEBUG oslo_vmware.api [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453899, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.363577] env[65385]: DEBUG nova.network.neutron [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Successfully created port: 299ec13e-c556-4a55-86fc-9c97363cc24d {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 884.385402] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39128c3e-a06f-4d58-922d-ab8ab2640ccc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.397292] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb43e6a5-35ad-4c3b-a811-cca06ef74550 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.432466] env[65385]: DEBUG nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 884.437533] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad97091-e465-4ca1-b114-8de344b846bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.442164] env[65385]: DEBUG nova.compute.manager [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Received event network-vif-plugged-ad2419b0-bd5b-4675-b73e-a9afc321bf4c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 884.442392] env[65385]: DEBUG oslo_concurrency.lockutils [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Acquiring lock "fa9a5278-1477-485a-9201-a37187488aef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.442650] env[65385]: DEBUG oslo_concurrency.lockutils [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Lock "fa9a5278-1477-485a-9201-a37187488aef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.442778] env[65385]: DEBUG oslo_concurrency.lockutils [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Lock "fa9a5278-1477-485a-9201-a37187488aef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 884.442974] env[65385]: DEBUG nova.compute.manager [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] No waiting events found dispatching network-vif-plugged-ad2419b0-bd5b-4675-b73e-a9afc321bf4c {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 884.443129] env[65385]: WARNING nova.compute.manager [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Received unexpected event network-vif-plugged-ad2419b0-bd5b-4675-b73e-a9afc321bf4c for instance with vm_state building and task_state spawning. [ 884.443322] env[65385]: DEBUG nova.compute.manager [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Received event network-changed-ad2419b0-bd5b-4675-b73e-a9afc321bf4c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 884.443497] env[65385]: DEBUG nova.compute.manager [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Refreshing instance network info cache due to event network-changed-ad2419b0-bd5b-4675-b73e-a9afc321bf4c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 884.443694] env[65385]: DEBUG oslo_concurrency.lockutils [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Acquiring lock "refresh_cache-fa9a5278-1477-485a-9201-a37187488aef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.443852] env[65385]: DEBUG oslo_concurrency.lockutils [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Acquired lock "refresh_cache-fa9a5278-1477-485a-9201-a37187488aef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 884.444019] env[65385]: DEBUG nova.network.neutron [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Refreshing network info cache for port ad2419b0-bd5b-4675-b73e-a9afc321bf4c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 884.455618] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b83315f-3a3b-42b6-b0d4-144fd82d271b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.473224] env[65385]: DEBUG nova.compute.provider_tree [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.524037] env[65385]: DEBUG oslo_vmware.api [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453897, 'name': PowerOffVM_Task, 'duration_secs': 0.329158} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.527822] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.527957] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.528244] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52acd029-c95b-01d3-b114-d536e1d07547, 'name': SearchDatastore_Task, 'duration_secs': 0.016828} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.529670] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8f3288d-e785-44b4-86fe-61a37647cf0f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.534546] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce03aaed-2d7b-4262-bf8c-6850171009d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.542186] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 884.542186] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529be963-7c2b-309d-fd14-25ba0f8b6924" [ 884.542186] env[65385]: _type = "Task" [ 884.542186] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.551936] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529be963-7c2b-309d-fd14-25ba0f8b6924, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.573196] env[65385]: DEBUG nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 884.594959] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453898, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.595429] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.596392] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f8a963-42db-4267-abe3-8b8c4b64855e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.609496] env[65385]: DEBUG oslo_vmware.api [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453899, 'name': PowerOffVM_Task, 'duration_secs': 0.255706} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.626831] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.626831] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.641717] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 6dae15ad-a340-47e3-ab6b-95b25043ca27/6dae15ad-a340-47e3-ab6b-95b25043ca27.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.644230] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6364aa7-ed5c-4cba-a7f8-27aaa390b828 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.646748] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a98e6ec-3169-4c87-8813-9f5ee4c25ad5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.663827] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.664070] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.664267] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleting the datastore file [datastore1] 42b9edf6-1873-49c3-8074-8eef654ac371 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.665634] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-696d1483-2a58-4a6a-8990-2238a25bb133 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.675300] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 884.675300] env[65385]: value = "task-4453902" [ 884.675300] env[65385]: _type = "Task" [ 884.675300] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.676610] env[65385]: DEBUG oslo_vmware.api [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for the task: (returnval){ [ 884.676610] env[65385]: value = "task-4453903" [ 884.676610] env[65385]: _type = "Task" [ 884.676610] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.691805] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453902, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.695564] env[65385]: DEBUG oslo_vmware.api [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.761832] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.762357] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.762654] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Deleting the datastore file [datastore2] af8d01e2-fb86-41c4-99a5-204b30eeda0c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.762976] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b8bbbf0-25cc-4aca-baf5-e02efc75d13b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.771573] env[65385]: DEBUG oslo_vmware.api [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for the task: (returnval){ [ 884.771573] env[65385]: value = "task-4453904" [ 884.771573] env[65385]: _type = "Task" [ 884.771573] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.780531] env[65385]: DEBUG oslo_vmware.api [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.949619] env[65385]: WARNING neutronclient.v2_0.client [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.950455] env[65385]: WARNING openstack [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.950710] env[65385]: WARNING openstack [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.977013] env[65385]: DEBUG nova.scheduler.client.report [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 885.038348] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.038596] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.038818] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.039046] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.039297] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.042061] env[65385]: INFO nova.compute.manager [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Terminating instance [ 885.060354] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529be963-7c2b-309d-fd14-25ba0f8b6924, 'name': SearchDatastore_Task, 'duration_secs': 0.013569} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.061266] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.061602] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] fa9a5278-1477-485a-9201-a37187488aef/fa9a5278-1477-485a-9201-a37187488aef.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 885.061888] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13c561eb-bc2f-47fc-bab2-bf558cd01510 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.072642] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 885.072642] env[65385]: value = "task-4453905" [ 885.072642] env[65385]: _type = "Task" [ 885.072642] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.087513] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.099629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.195575] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453902, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.196254] env[65385]: DEBUG oslo_vmware.api [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.284712] env[65385]: DEBUG oslo_vmware.api [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.459511] env[65385]: DEBUG nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 885.485200] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.085s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.488037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.012s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.489027] env[65385]: DEBUG nova.objects.instance [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lazy-loading 'resources' on Instance uuid 75ba706f-2e72-4f84-b02f-db4381951e77 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.497138] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 885.497409] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 885.497639] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 885.497873] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 885.498008] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 885.498482] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 885.498482] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.498647] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 885.498691] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 885.498874] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 885.499110] env[65385]: DEBUG nova.virt.hardware [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 885.500141] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4966a24-846e-406f-a10b-0819f097abe7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.512093] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b233042-41f5-49c8-85bf-23f31118c7bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.531721] env[65385]: INFO nova.scheduler.client.report [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Deleted allocations for instance 7fcbdc79-688c-479f-94e2-f4542abe714b [ 885.551966] env[65385]: DEBUG nova.compute.manager [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 885.552262] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.553461] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0c16f8-c222-4e27-bce7-7fd659832ada {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.562630] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.563058] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d27f486c-9b1d-43b5-8e33-2b17113372c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.574224] env[65385]: DEBUG oslo_vmware.api [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 885.574224] env[65385]: value = "task-4453906" [ 885.574224] env[65385]: _type = "Task" [ 885.574224] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.599986] env[65385]: DEBUG oslo_vmware.api [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.600390] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453905, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.696226] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453902, 'name': ReconfigVM_Task, 'duration_secs': 0.945716} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.701067] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 6dae15ad-a340-47e3-ab6b-95b25043ca27/6dae15ad-a340-47e3-ab6b-95b25043ca27.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.701067] env[65385]: DEBUG oslo_vmware.api [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Task: {'id': task-4453903, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.705963} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.701067] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2fc81a99-edc0-4a03-a625-b92bf84663b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.702926] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.703204] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.703327] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.703506] env[65385]: INFO nova.compute.manager [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Took 1.72 seconds to destroy the instance on the hypervisor. [ 885.703752] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 885.703964] env[65385]: DEBUG nova.compute.manager [-] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 885.704072] env[65385]: DEBUG nova.network.neutron [-] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 885.704333] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.704852] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.705232] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.719984] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 885.719984] env[65385]: value = "task-4453907" [ 885.719984] env[65385]: _type = "Task" [ 885.719984] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.730882] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453907, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.785774] env[65385]: DEBUG oslo_vmware.api [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Task: {'id': task-4453904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.659419} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.786076] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.786266] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.786447] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.786639] env[65385]: INFO nova.compute.manager [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Took 1.73 seconds to destroy the instance on the hypervisor. [ 885.786893] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 885.787114] env[65385]: DEBUG nova.compute.manager [-] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 885.787211] env[65385]: DEBUG nova.network.neutron [-] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 885.787533] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.788072] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.788367] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.033783] env[65385]: DEBUG nova.network.neutron [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Successfully updated port: 299ec13e-c556-4a55-86fc-9c97363cc24d {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 886.044906] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00cee9bc-12c8-4e38-9505-7647e827563d tempest-FloatingIPsAssociationTestJSON-566088622 tempest-FloatingIPsAssociationTestJSON-566088622-project-member] Lock "7fcbdc79-688c-479f-94e2-f4542abe714b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.022s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.047075] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.047905] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquired lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.047905] env[65385]: DEBUG nova.network.neutron [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 886.065585] env[65385]: WARNING openstack [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.066316] env[65385]: WARNING openstack [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.096736] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.843665} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.100223] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] fa9a5278-1477-485a-9201-a37187488aef/fa9a5278-1477-485a-9201-a37187488aef.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.100480] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.100813] env[65385]: DEBUG oslo_vmware.api [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453906, 'name': PowerOffVM_Task, 'duration_secs': 0.424342} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.101628] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-529cac82-a720-4cce-aaec-43abcad4093b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.103829] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.104024] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.104578] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0851388f-cf88-4eaf-af87-a152e294485b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.119871] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 886.119871] env[65385]: value = "task-4453908" [ 886.119871] env[65385]: _type = "Task" [ 886.119871] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.131431] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453908, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.153951] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.160384] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.198032] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.198032] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.198032] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleting the datastore file [datastore1] e02e3848-1b1e-426b-bc0f-24c8e232a9db {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.198032] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64423261-60f4-4418-ae23-fb49f3d98663 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.206239] env[65385]: DEBUG oslo_vmware.api [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for the task: (returnval){ [ 886.206239] env[65385]: value = "task-4453910" [ 886.206239] env[65385]: _type = "Task" [ 886.206239] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.215734] env[65385]: DEBUG oslo_vmware.api [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453910, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.230299] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453907, 'name': Rename_Task, 'duration_secs': 0.283829} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.231348] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.231348] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a1549c4-c9e3-4a33-a929-f9937bf5a4f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.239074] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 886.239074] env[65385]: value = "task-4453911" [ 886.239074] env[65385]: _type = "Task" [ 886.239074] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.249959] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.466350] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2898300a-f49e-4a73-a204-58ef26604043 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.477408] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d975c3c9-42bc-4984-9b64-042de14119fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.518474] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ae2063-033b-4806-ba2f-dc1349a1f7aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.523017] env[65385]: WARNING neutronclient.v2_0.client [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.523687] env[65385]: WARNING openstack [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.524044] env[65385]: WARNING openstack [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.541265] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93918959-0519-4fb9-9053-e2dd0fd4cbf0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.558866] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.559486] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.569284] env[65385]: DEBUG nova.compute.provider_tree [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.632981] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453908, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.243861} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.633247] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.634189] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb7a8c1-7a44-4f0d-905b-cd939f5115fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.661703] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] fa9a5278-1477-485a-9201-a37187488aef/fa9a5278-1477-485a-9201-a37187488aef.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.663125] env[65385]: DEBUG nova.network.neutron [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 886.665960] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b37fa085-8b20-4f7f-8045-f7016feddde0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.690182] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 886.690182] env[65385]: value = "task-4453912" [ 886.690182] env[65385]: _type = "Task" [ 886.690182] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.697886] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453912, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.719558] env[65385]: DEBUG oslo_vmware.api [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Task: {'id': task-4453910, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.44159} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.720587] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.720587] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.720587] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.720587] env[65385]: INFO nova.compute.manager [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Took 1.17 seconds to destroy the instance on the hypervisor. [ 886.721280] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 886.721280] env[65385]: DEBUG nova.compute.manager [-] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 886.721280] env[65385]: DEBUG nova.network.neutron [-] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 886.721280] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.722555] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.722555] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.732079] env[65385]: DEBUG nova.network.neutron [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Updated VIF entry in instance network info cache for port ad2419b0-bd5b-4675-b73e-a9afc321bf4c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 886.732483] env[65385]: DEBUG nova.network.neutron [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Updating instance_info_cache with network_info: [{"id": "ad2419b0-bd5b-4675-b73e-a9afc321bf4c", "address": "fa:16:3e:f7:99:c9", "network": {"id": "5da125f0-44fc-462c-8a94-96a6f22a3a68", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-467202416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21234ef9798c48278f3f65cdd9913f3c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad2419b0-bd", "ovs_interfaceid": "ad2419b0-bd5b-4675-b73e-a9afc321bf4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.747652] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.748065] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.760743] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453911, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.793852] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.965444] env[65385]: DEBUG nova.network.neutron [-] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.976823] env[65385]: DEBUG nova.network.neutron [-] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 887.072614] env[65385]: DEBUG nova.scheduler.client.report [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 887.081688] env[65385]: WARNING neutronclient.v2_0.client [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.082373] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.082765] env[65385]: WARNING openstack [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.201024] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.201234] env[65385]: DEBUG nova.network.neutron [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Updating instance_info_cache with network_info: [{"id": "299ec13e-c556-4a55-86fc-9c97363cc24d", "address": "fa:16:3e:8b:c0:4d", "network": {"id": "adf4395d-70f5-4db9-9730-2fd88ca189a9", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-307667077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e1187e438e194c4587f8824110dc23f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299ec13e-c5", "ovs_interfaceid": "299ec13e-c556-4a55-86fc-9c97363cc24d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 887.235166] env[65385]: DEBUG oslo_concurrency.lockutils [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] Releasing lock "refresh_cache-fa9a5278-1477-485a-9201-a37187488aef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.236656] env[65385]: DEBUG nova.compute.manager [req-5b93a029-5080-4685-a026-e5725972fc62 req-b10c6306-948c-433a-bebf-7bccbdd9478c service nova] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Received event network-vif-deleted-3f2342b2-bc3c-489f-a61b-ffb18a20cce6 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 887.249757] env[65385]: DEBUG oslo_vmware.api [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453911, 'name': PowerOnVM_Task, 'duration_secs': 0.852638} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.250393] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.250786] env[65385]: INFO nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Took 10.30 seconds to spawn the instance on the hypervisor. [ 887.250786] env[65385]: DEBUG nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 887.251577] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eead0ea-63b6-4ce7-bb91-64e3cd0b90cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.468174] env[65385]: INFO nova.compute.manager [-] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Took 1.68 seconds to deallocate network for instance. [ 887.480398] env[65385]: INFO nova.compute.manager [-] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Took 1.78 seconds to deallocate network for instance. [ 887.553835] env[65385]: DEBUG nova.network.neutron [-] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 887.565544] env[65385]: DEBUG nova.compute.manager [req-b33bcddf-ba07-4519-83f5-5bdad9a4b311 req-a46aef7f-ddfe-4eb3-8785-1b49c5be5dda service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Received event network-vif-plugged-299ec13e-c556-4a55-86fc-9c97363cc24d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 887.565729] env[65385]: DEBUG oslo_concurrency.lockutils [req-b33bcddf-ba07-4519-83f5-5bdad9a4b311 req-a46aef7f-ddfe-4eb3-8785-1b49c5be5dda service nova] Acquiring lock "cf37e10a-14f4-4456-bad0-d7528457658d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 887.565967] env[65385]: DEBUG oslo_concurrency.lockutils [req-b33bcddf-ba07-4519-83f5-5bdad9a4b311 req-a46aef7f-ddfe-4eb3-8785-1b49c5be5dda service nova] Lock "cf37e10a-14f4-4456-bad0-d7528457658d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.566157] env[65385]: DEBUG oslo_concurrency.lockutils [req-b33bcddf-ba07-4519-83f5-5bdad9a4b311 req-a46aef7f-ddfe-4eb3-8785-1b49c5be5dda service nova] Lock "cf37e10a-14f4-4456-bad0-d7528457658d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.566385] env[65385]: DEBUG nova.compute.manager [req-b33bcddf-ba07-4519-83f5-5bdad9a4b311 req-a46aef7f-ddfe-4eb3-8785-1b49c5be5dda service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] No waiting events found dispatching network-vif-plugged-299ec13e-c556-4a55-86fc-9c97363cc24d {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 887.566517] env[65385]: WARNING nova.compute.manager [req-b33bcddf-ba07-4519-83f5-5bdad9a4b311 req-a46aef7f-ddfe-4eb3-8785-1b49c5be5dda service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Received unexpected event network-vif-plugged-299ec13e-c556-4a55-86fc-9c97363cc24d for instance with vm_state building and task_state spawning. [ 887.584018] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.094s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.585702] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.838s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.586081] env[65385]: DEBUG nova.objects.instance [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lazy-loading 'resources' on Instance uuid 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.598118] env[65385]: DEBUG nova.compute.manager [req-c43eb296-bf08-4af8-9002-019a4f6fe22b req-dcaa8b14-a47b-4e0f-bc13-49015ec4eb67 service nova] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Received event network-vif-deleted-03771ba6-a616-4715-afb9-19306b1b0903 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 887.607753] env[65385]: INFO nova.scheduler.client.report [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted allocations for instance 75ba706f-2e72-4f84-b02f-db4381951e77 [ 887.698149] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453912, 'name': ReconfigVM_Task, 'duration_secs': 0.683722} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.698484] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Reconfigured VM instance instance-00000046 to attach disk [datastore2] fa9a5278-1477-485a-9201-a37187488aef/fa9a5278-1477-485a-9201-a37187488aef.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.699242] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-258c0571-8057-4b5c-a792-6e3b5da1241f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.707293] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Releasing lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.707751] env[65385]: DEBUG nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Instance network_info: |[{"id": "299ec13e-c556-4a55-86fc-9c97363cc24d", "address": "fa:16:3e:8b:c0:4d", "network": {"id": "adf4395d-70f5-4db9-9730-2fd88ca189a9", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-307667077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e1187e438e194c4587f8824110dc23f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299ec13e-c5", "ovs_interfaceid": "299ec13e-c556-4a55-86fc-9c97363cc24d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 887.708176] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 887.708176] env[65385]: value = "task-4453913" [ 887.708176] env[65385]: _type = "Task" [ 887.708176] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.708520] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:c0:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43ad01d2-c7dd-453c-a929-8ad76294d13c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '299ec13e-c556-4a55-86fc-9c97363cc24d', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.716508] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Creating folder: Project (e1187e438e194c4587f8824110dc23f6). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 887.717240] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-49c103a6-29e3-485d-97a7-987afde257f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.728622] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453913, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.734022] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Created folder: Project (e1187e438e194c4587f8824110dc23f6) in parent group-v870881. [ 887.734022] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Creating folder: Instances. Parent ref: group-v871086. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 887.734022] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83bf0a46-c3f8-49a3-b4e3-2dfa7bf2e334 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.742297] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Created folder: Instances in parent group-v871086. [ 887.742530] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 887.742726] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.742938] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce705806-2d6c-454e-a983-d36fd4ad53fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.768273] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.768273] env[65385]: value = "task-4453916" [ 887.768273] env[65385]: _type = "Task" [ 887.768273] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.774049] env[65385]: INFO nova.compute.manager [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Took 32.28 seconds to build instance. [ 887.781751] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453916, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.976226] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 887.987930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 888.060060] env[65385]: INFO nova.compute.manager [-] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Took 1.34 seconds to deallocate network for instance. [ 888.117176] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4fa9e699-06da-466b-80ae-aed197a49100 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "75ba706f-2e72-4f84-b02f-db4381951e77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.363s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.229026] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453913, 'name': Rename_Task, 'duration_secs': 0.258951} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.232289] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.232898] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b17b7a1d-015b-4661-a4d1-8777d9f57393 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.241817] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 888.241817] env[65385]: value = "task-4453917" [ 888.241817] env[65385]: _type = "Task" [ 888.241817] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.256026] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453917, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.277071] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5208aeba-bb2b-4c64-b7bb-b11fb1875c79 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.784s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.287785] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453916, 'name': CreateVM_Task, 'duration_secs': 0.356898} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.288728] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.288728] env[65385]: WARNING neutronclient.v2_0.client [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.290106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.290106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.290106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 888.290386] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30537de6-80c0-4bc6-87ce-cff7fd0b62c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.300068] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 888.300068] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5278ba73-92e5-6907-2513-e3d740c863c5" [ 888.300068] env[65385]: _type = "Task" [ 888.300068] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.310309] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5278ba73-92e5-6907-2513-e3d740c863c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.438898] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2039435-c49c-4ff9-b4ad-0d4aec51106a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.451097] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15160a05-491a-4997-b297-2dbbf996165c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.488842] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3000b14-c943-47ca-bd0c-5f6f26e800ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.495352] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c427d5-79c4-41c0-bed7-764432165ad9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.511706] env[65385]: DEBUG nova.compute.provider_tree [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.566667] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 888.753107] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453917, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.819524] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5278ba73-92e5-6907-2513-e3d740c863c5, 'name': SearchDatastore_Task, 'duration_secs': 0.016664} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.822048] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.822048] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.822048] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.822048] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.822048] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.822048] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cad0b01-83df-43ce-bd68-1bbe93fc1707 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.831395] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.831588] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.832435] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7af2123-9b36-4bb9-9369-64d8a135fa46 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.839398] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 888.839398] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fdfd0-ff7d-5c42-4e4d-acf82b215c67" [ 888.839398] env[65385]: _type = "Task" [ 888.839398] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.855900] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fdfd0-ff7d-5c42-4e4d-acf82b215c67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.017459] env[65385]: DEBUG nova.scheduler.client.report [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.254030] env[65385]: DEBUG oslo_vmware.api [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453917, 'name': PowerOnVM_Task, 'duration_secs': 0.763039} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.254523] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.254523] env[65385]: INFO nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Took 9.77 seconds to spawn the instance on the hypervisor. [ 889.254932] env[65385]: DEBUG nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 889.255679] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5428c1f9-2714-4a6a-a78a-c25c463e2a4d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.351087] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fdfd0-ff7d-5c42-4e4d-acf82b215c67, 'name': SearchDatastore_Task, 'duration_secs': 0.017449} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.351905] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf91e636-0b53-44af-922b-f78eafc5705b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.357970] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 889.357970] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6e975-717b-a01a-8eed-005f134014e7" [ 889.357970] env[65385]: _type = "Task" [ 889.357970] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.367548] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6e975-717b-a01a-8eed-005f134014e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.526915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.940s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.530031] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 22.574s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.562767] env[65385]: INFO nova.scheduler.client.report [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleted allocations for instance 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f [ 889.780368] env[65385]: INFO nova.compute.manager [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Took 31.84 seconds to build instance. [ 889.876028] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a6e975-717b-a01a-8eed-005f134014e7, 'name': SearchDatastore_Task, 'duration_secs': 0.016734} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.876028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.876028] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] cf37e10a-14f4-4456-bad0-d7528457658d/cf37e10a-14f4-4456-bad0-d7528457658d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.876403] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b032387-e354-4310-ae27-d057ec8bc559 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.885964] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 889.885964] env[65385]: value = "task-4453918" [ 889.885964] env[65385]: _type = "Task" [ 889.885964] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.896792] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.043153] env[65385]: DEBUG nova.compute.manager [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Received event network-changed-299ec13e-c556-4a55-86fc-9c97363cc24d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 890.043277] env[65385]: DEBUG nova.compute.manager [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Refreshing instance network info cache due to event network-changed-299ec13e-c556-4a55-86fc-9c97363cc24d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 890.043485] env[65385]: DEBUG oslo_concurrency.lockutils [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Acquiring lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.043619] env[65385]: DEBUG oslo_concurrency.lockutils [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Acquired lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.044013] env[65385]: DEBUG nova.network.neutron [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Refreshing network info cache for port 299ec13e-c556-4a55-86fc-9c97363cc24d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 890.077526] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e52f8289-b0ac-4ad6-a2d7-eefc766ad7c7 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "703ed9b6-7cd8-4a84-9847-d34fb1c51a3f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.903s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.290721] env[65385]: DEBUG oslo_concurrency.lockutils [None req-64c682c8-25fa-4b89-9883-8fe368d99d54 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "fa9a5278-1477-485a-9201-a37187488aef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.367s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.399915] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453918, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.425281] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80e0db8-37d2-447a-ae21-35413719e590 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.436544] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4e7bb0-aac2-4964-8be7-98d22c14b3ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.484468] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df1ef61-8451-4662-b781-16d5961ede84 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.496472] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9c0ec5-66e4-4671-9a7f-3e48170aaaac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.520964] env[65385]: DEBUG nova.compute.provider_tree [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.547560] env[65385]: WARNING neutronclient.v2_0.client [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.548708] env[65385]: WARNING openstack [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.550378] env[65385]: WARNING openstack [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.639639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "eb9d0510-b453-4695-9e1d-731217b9f8ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.640059] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.643020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "eb9d0510-b453-4695-9e1d-731217b9f8ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.643020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.643020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.644368] env[65385]: INFO nova.compute.manager [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Terminating instance [ 890.855136] env[65385]: WARNING openstack [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.855136] env[65385]: WARNING openstack [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.899544] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453918, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.940227} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.904231] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] cf37e10a-14f4-4456-bad0-d7528457658d/cf37e10a-14f4-4456-bad0-d7528457658d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.904455] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.904914] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dcab1aa9-dccc-489d-89c2-01cfd70226d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.912503] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 890.912503] env[65385]: value = "task-4453919" [ 890.912503] env[65385]: _type = "Task" [ 890.912503] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.929140] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453919, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.932130] env[65385]: WARNING neutronclient.v2_0.client [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.932797] env[65385]: WARNING openstack [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.933161] env[65385]: WARNING openstack [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.025013] env[65385]: DEBUG nova.scheduler.client.report [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 891.072821] env[65385]: DEBUG nova.network.neutron [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Updated VIF entry in instance network info cache for port 299ec13e-c556-4a55-86fc-9c97363cc24d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 891.072919] env[65385]: DEBUG nova.network.neutron [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Updating instance_info_cache with network_info: [{"id": "299ec13e-c556-4a55-86fc-9c97363cc24d", "address": "fa:16:3e:8b:c0:4d", "network": {"id": "adf4395d-70f5-4db9-9730-2fd88ca189a9", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-307667077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e1187e438e194c4587f8824110dc23f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299ec13e-c5", "ovs_interfaceid": "299ec13e-c556-4a55-86fc-9c97363cc24d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 891.150362] env[65385]: DEBUG nova.compute.manager [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 891.150567] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.152172] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4868d512-f559-41aa-b763-5cdb23d07bd9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.161742] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.161742] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42e9dd25-a57d-42c2-8f05-3dcff9937e59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.169087] env[65385]: DEBUG oslo_vmware.api [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 891.169087] env[65385]: value = "task-4453920" [ 891.169087] env[65385]: _type = "Task" [ 891.169087] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.179741] env[65385]: DEBUG oslo_vmware.api [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.245156] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.245932] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.424972] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453919, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.250926} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.426377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.426377] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4f7fb3-0e7f-4131-b05f-1b8ab1186724 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.464703] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] cf37e10a-14f4-4456-bad0-d7528457658d/cf37e10a-14f4-4456-bad0-d7528457658d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.465072] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9486769d-98e3-4ef9-a81c-928fb78b2126 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.491201] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 891.491201] env[65385]: value = "task-4453921" [ 891.491201] env[65385]: _type = "Task" [ 891.491201] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.507689] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453921, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.575276] env[65385]: DEBUG oslo_concurrency.lockutils [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] Releasing lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.576732] env[65385]: DEBUG nova.compute.manager [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Received event network-vif-deleted-e615249d-2b37-4211-abe1-70015ea3d4cf {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 891.576732] env[65385]: DEBUG nova.compute.manager [req-6204fd53-993f-4011-bfd2-065b83ac4045 req-1a417011-14e9-47a0-af50-92a85a07a7cf service nova] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Received event network-vif-deleted-2e9cf308-c94b-4a76-9258-93c5d7563146 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 891.682336] env[65385]: DEBUG oslo_vmware.api [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453920, 'name': PowerOffVM_Task, 'duration_secs': 0.419509} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.682638] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.682813] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.683291] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-caa47add-be69-42c4-a4d5-ef24dc99563a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.750408] env[65385]: DEBUG nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 891.763037] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.763702] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.763702] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleting the datastore file [datastore2] eb9d0510-b453-4695-9e1d-731217b9f8ec {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.764739] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4578943b-d7aa-4f86-9dfe-ce9b87f239b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.773257] env[65385]: DEBUG oslo_vmware.api [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for the task: (returnval){ [ 891.773257] env[65385]: value = "task-4453923" [ 891.773257] env[65385]: _type = "Task" [ 891.773257] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.790408] env[65385]: DEBUG oslo_vmware.api [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.005932] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453921, 'name': ReconfigVM_Task, 'duration_secs': 0.310001} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.007276] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Reconfigured VM instance instance-00000047 to attach disk [datastore2] cf37e10a-14f4-4456-bad0-d7528457658d/cf37e10a-14f4-4456-bad0-d7528457658d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.007616] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8be20a79-7f2a-4680-a02b-6d3aeb3ecc4c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.016443] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 892.016443] env[65385]: value = "task-4453924" [ 892.016443] env[65385]: _type = "Task" [ 892.016443] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.026522] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453924, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.046317] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.518s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.046701] env[65385]: DEBUG nova.compute.manager [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=65385) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5416}} [ 892.050513] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.484s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 892.052883] env[65385]: DEBUG nova.objects.instance [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lazy-loading 'resources' on Instance uuid 65f9bc55-7bab-43b0-a974-eb5080389b7d {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.292302] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 892.300222] env[65385]: DEBUG oslo_vmware.api [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Task: {'id': task-4453923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.47845} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.300222] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.300515] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.300805] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.301114] env[65385]: INFO nova.compute.manager [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Took 1.15 seconds to destroy the instance on the hypervisor. [ 892.301473] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 892.301826] env[65385]: DEBUG nova.compute.manager [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 892.302064] env[65385]: DEBUG nova.network.neutron [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 892.302447] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.303200] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.303632] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.351992] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.532044] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453924, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.626354] env[65385]: INFO nova.scheduler.client.report [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted allocation for migration 73856ca4-6bf7-447d-9523-fd6a10d8cb2a [ 892.837352] env[65385]: DEBUG nova.compute.manager [req-e77b53b2-0f28-42c1-9016-376353f08454 req-571dd4f0-a416-47d9-a03d-ffa68cf9552a service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Received event network-vif-deleted-2206d06f-505b-4211-9823-29d900820228 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 892.837549] env[65385]: INFO nova.compute.manager [req-e77b53b2-0f28-42c1-9016-376353f08454 req-571dd4f0-a416-47d9-a03d-ffa68cf9552a service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Neutron deleted interface 2206d06f-505b-4211-9823-29d900820228; detaching it from the instance and deleting it from the info cache [ 892.837715] env[65385]: DEBUG nova.network.neutron [req-e77b53b2-0f28-42c1-9016-376353f08454 req-571dd4f0-a416-47d9-a03d-ffa68cf9552a service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 892.989127] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc7dcea-1ff5-46e4-9fc6-627daf617d7e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.000314] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c980f5-a307-482f-ae97-f0f971dac808 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.041444] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83220b29-1be6-4e5a-ac01-7efbcf1b59e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.058459] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b9c884-48c4-4ce3-995e-bb266fb3e2a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.063352] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453924, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.076211] env[65385]: DEBUG nova.compute.provider_tree [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.140353] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6dafca4c-6549-4077-9ab6-484fd38bb64a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 29.503s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.166445] env[65385]: DEBUG nova.network.neutron [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 893.341586] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7b00b476-e5d9-45d1-b102-c1ffd2db0461 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.356172] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681bc74b-b868-48ec-90c8-449c72aad032 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.391196] env[65385]: DEBUG nova.compute.manager [req-e77b53b2-0f28-42c1-9016-376353f08454 req-571dd4f0-a416-47d9-a03d-ffa68cf9552a service nova] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Detach interface failed, port_id=2206d06f-505b-4211-9823-29d900820228, reason: Instance eb9d0510-b453-4695-9e1d-731217b9f8ec could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 893.548301] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453924, 'name': Rename_Task, 'duration_secs': 1.173549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.548721] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.549668] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3bcd61ec-c717-4b65-9cb1-122fc597e91b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.551668] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "fa9a5278-1477-485a-9201-a37187488aef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 893.551668] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "fa9a5278-1477-485a-9201-a37187488aef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 893.551668] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "fa9a5278-1477-485a-9201-a37187488aef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 893.551668] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "fa9a5278-1477-485a-9201-a37187488aef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 893.551823] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "fa9a5278-1477-485a-9201-a37187488aef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.554977] env[65385]: INFO nova.compute.manager [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Terminating instance [ 893.560433] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 893.560433] env[65385]: value = "task-4453925" [ 893.560433] env[65385]: _type = "Task" [ 893.560433] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.569887] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453925, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.579608] env[65385]: DEBUG nova.scheduler.client.report [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.669655] env[65385]: INFO nova.compute.manager [-] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Took 1.37 seconds to deallocate network for instance. [ 894.059353] env[65385]: DEBUG nova.compute.manager [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 894.059594] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.060739] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99c496f-bb40-43b6-8a58-61e65d247b5e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.078104] env[65385]: DEBUG oslo_vmware.api [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453925, 'name': PowerOnVM_Task, 'duration_secs': 0.503866} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.078395] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.078660] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.078868] env[65385]: INFO nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Took 8.62 seconds to spawn the instance on the hypervisor. [ 894.079449] env[65385]: DEBUG nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 894.079449] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2da400f4-5535-4f1e-9cb7-83af5bb5a97c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.081770] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a5745d-0f33-411b-b27b-c9a2303f8ed9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.084630] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.034s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.087442] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.996s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.087683] env[65385]: DEBUG nova.objects.instance [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lazy-loading 'resources' on Instance uuid 9908d395-545d-4caf-9757-320d6253d61e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.093653] env[65385]: DEBUG oslo_vmware.api [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 894.093653] env[65385]: value = "task-4453926" [ 894.093653] env[65385]: _type = "Task" [ 894.093653] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.114093] env[65385]: DEBUG oslo_vmware.api [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.116562] env[65385]: INFO nova.scheduler.client.report [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Deleted allocations for instance 65f9bc55-7bab-43b0-a974-eb5080389b7d [ 894.171089] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "6dae15ad-a340-47e3-ab6b-95b25043ca27" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.171405] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.172037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "6dae15ad-a340-47e3-ab6b-95b25043ca27-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.172037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.172037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.174940] env[65385]: INFO nova.compute.manager [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Terminating instance [ 894.178800] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.368758] env[65385]: DEBUG nova.objects.instance [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'flavor' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.612036] env[65385]: DEBUG oslo_vmware.api [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453926, 'name': PowerOffVM_Task, 'duration_secs': 0.337549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.612980] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.613176] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.613433] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b427395-e4c7-4b30-be4a-e72b145a0562 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.615842] env[65385]: INFO nova.compute.manager [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Took 31.14 seconds to build instance. [ 894.628543] env[65385]: DEBUG oslo_concurrency.lockutils [None req-573af72c-9574-44d3-8cc6-3c0ff9842027 tempest-ServerShowV247Test-400054191 tempest-ServerShowV247Test-400054191-project-member] Lock "65f9bc55-7bab-43b0-a974-eb5080389b7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.580s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.678759] env[65385]: DEBUG nova.compute.manager [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 894.679078] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.680228] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1491bb-56cf-47c5-80b3-c940e4a1e89f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.690039] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.690240] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a115dc4-5239-4db7-8715-b54792c327cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.699162] env[65385]: DEBUG oslo_vmware.api [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 894.699162] env[65385]: value = "task-4453928" [ 894.699162] env[65385]: _type = "Task" [ 894.699162] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.708496] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 894.708496] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 894.708698] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Deleting the datastore file [datastore2] fa9a5278-1477-485a-9201-a37187488aef {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.708945] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2917e88-1c8b-4998-b1c1-1ef9e805b892 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.718097] env[65385]: DEBUG oslo_vmware.api [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453928, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.725231] env[65385]: DEBUG oslo_vmware.api [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for the task: (returnval){ [ 894.725231] env[65385]: value = "task-4453929" [ 894.725231] env[65385]: _type = "Task" [ 894.725231] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.736561] env[65385]: DEBUG oslo_vmware.api [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.876683] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.876917] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.877118] env[65385]: DEBUG nova.network.neutron [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 894.877299] env[65385]: DEBUG nova.objects.instance [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'info_cache' on Instance uuid 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.057854] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "3febf6af-699a-4fa1-b079-3790cd1095e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.058107] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.085268] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af98ed2-6495-4aaa-8955-2673cd2d1492 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.097844] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c4acb3-052a-4a29-805f-b2fa4d731b14 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.132999] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6faf1fdf-9abd-4330-b653-f4c863fa198a tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.675s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 895.134891] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74043d67-0eca-4565-83dc-255af992b1fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.144237] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f687a71-0935-4b6b-a073-ba9dc208e5e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.160942] env[65385]: DEBUG nova.compute.provider_tree [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.213354] env[65385]: DEBUG oslo_vmware.api [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453928, 'name': PowerOffVM_Task, 'duration_secs': 0.265152} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.214262] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.214262] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.214262] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5bcd38c-617a-40ef-aedd-8759700d380d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.236879] env[65385]: DEBUG oslo_vmware.api [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Task: {'id': task-4453929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197226} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.237307] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.237516] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.237711] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.238378] env[65385]: INFO nova.compute.manager [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] [instance: fa9a5278-1477-485a-9201-a37187488aef] Took 1.18 seconds to destroy the instance on the hypervisor. [ 895.238378] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 895.239726] env[65385]: DEBUG nova.compute.manager [-] [instance: fa9a5278-1477-485a-9201-a37187488aef] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 895.239904] env[65385]: DEBUG nova.network.neutron [-] [instance: fa9a5278-1477-485a-9201-a37187488aef] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 895.240178] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.242418] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.242418] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.290462] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.290627] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.290809] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Deleting the datastore file [datastore1] 6dae15ad-a340-47e3-ab6b-95b25043ca27 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.292098] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.294188] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b94a1dc5-4c41-4106-aa55-cc131fe0d788 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.304069] env[65385]: DEBUG oslo_vmware.api [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for the task: (returnval){ [ 895.304069] env[65385]: value = "task-4453931" [ 895.304069] env[65385]: _type = "Task" [ 895.304069] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.313752] env[65385]: DEBUG oslo_vmware.api [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.388285] env[65385]: DEBUG nova.objects.base [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Object Instance<89f3c815-8671-47ce-9e74-bf6e652bb3c3> lazy-loaded attributes: flavor,info_cache {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 895.563065] env[65385]: DEBUG nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 895.665019] env[65385]: DEBUG nova.scheduler.client.report [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 895.699332] env[65385]: DEBUG nova.compute.manager [req-ed04d579-d686-4d72-9fd1-a35a9602b51a req-35b65e14-02da-43c7-8f22-3a5cd4c396e2 service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Received event network-vif-deleted-ad2419b0-bd5b-4675-b73e-a9afc321bf4c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 895.699468] env[65385]: INFO nova.compute.manager [req-ed04d579-d686-4d72-9fd1-a35a9602b51a req-35b65e14-02da-43c7-8f22-3a5cd4c396e2 service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Neutron deleted interface ad2419b0-bd5b-4675-b73e-a9afc321bf4c; detaching it from the instance and deleting it from the info cache [ 895.699594] env[65385]: DEBUG nova.network.neutron [req-ed04d579-d686-4d72-9fd1-a35a9602b51a req-35b65e14-02da-43c7-8f22-3a5cd4c396e2 service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 895.754577] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "cf37e10a-14f4-4456-bad0-d7528457658d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.754835] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.755087] env[65385]: INFO nova.compute.manager [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Rebooting instance [ 895.814720] env[65385]: DEBUG oslo_vmware.api [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Task: {'id': task-4453931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165215} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.814878] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.815163] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.815318] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.816499] env[65385]: INFO nova.compute.manager [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Took 1.14 seconds to destroy the instance on the hypervisor. [ 895.816499] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 895.816499] env[65385]: DEBUG nova.compute.manager [-] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 895.816499] env[65385]: DEBUG nova.network.neutron [-] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 895.816881] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.817320] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.817579] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.873051] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.894349] env[65385]: WARNING neutronclient.v2_0.client [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.896037] env[65385]: WARNING openstack [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.896586] env[65385]: WARNING openstack [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.066156] env[65385]: WARNING openstack [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.066156] env[65385]: WARNING openstack [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.104248] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 896.122583] env[65385]: DEBUG nova.network.neutron [-] [instance: fa9a5278-1477-485a-9201-a37187488aef] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 896.173323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.176666] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.877s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.178777] env[65385]: INFO nova.compute.claims [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.183790] env[65385]: WARNING neutronclient.v2_0.client [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 896.184433] env[65385]: WARNING openstack [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.184782] env[65385]: WARNING openstack [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.203261] env[65385]: INFO nova.scheduler.client.report [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Deleted allocations for instance 9908d395-545d-4caf-9757-320d6253d61e [ 896.204363] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1ee488c-2d49-467c-9590-94f8ffc97398 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.218923] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c36fe9-d1e2-47ac-adb0-05987143cd62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.261377] env[65385]: DEBUG nova.compute.manager [req-ed04d579-d686-4d72-9fd1-a35a9602b51a req-35b65e14-02da-43c7-8f22-3a5cd4c396e2 service nova] [instance: fa9a5278-1477-485a-9201-a37187488aef] Detach interface failed, port_id=ad2419b0-bd5b-4675-b73e-a9afc321bf4c, reason: Instance fa9a5278-1477-485a-9201-a37187488aef could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 896.282217] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.282364] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquired lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 896.282634] env[65385]: DEBUG nova.network.neutron [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 896.310495] env[65385]: DEBUG nova.network.neutron [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [{"id": "18c18d61-0493-40e3-8883-c90faabf147b", "address": "fa:16:3e:d1:d7:6e", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18c18d61-04", "ovs_interfaceid": "18c18d61-0493-40e3-8883-c90faabf147b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 896.615667] env[65385]: DEBUG nova.network.neutron [-] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 896.625839] env[65385]: INFO nova.compute.manager [-] [instance: fa9a5278-1477-485a-9201-a37187488aef] Took 1.39 seconds to deallocate network for instance. [ 896.714992] env[65385]: DEBUG oslo_concurrency.lockutils [None req-476e7883-ded2-4fac-ac93-a853155233ac tempest-ServersNegativeTestMultiTenantJSON-1026424368 tempest-ServersNegativeTestMultiTenantJSON-1026424368-project-member] Lock "9908d395-545d-4caf-9757-320d6253d61e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.566s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.785774] env[65385]: WARNING neutronclient.v2_0.client [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 896.787033] env[65385]: WARNING openstack [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.787033] env[65385]: WARNING openstack [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.813914] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-89f3c815-8671-47ce-9e74-bf6e652bb3c3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 896.971684] env[65385]: WARNING openstack [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.972181] env[65385]: WARNING openstack [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.047394] env[65385]: WARNING neutronclient.v2_0.client [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.048051] env[65385]: WARNING openstack [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.048416] env[65385]: WARNING openstack [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.066675] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "354527f0-007d-449f-9e15-48ce1d91876f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.066883] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "354527f0-007d-449f-9e15-48ce1d91876f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 897.119193] env[65385]: INFO nova.compute.manager [-] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Took 1.30 seconds to deallocate network for instance. [ 897.133839] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.187947] env[65385]: DEBUG nova.network.neutron [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Updating instance_info_cache with network_info: [{"id": "299ec13e-c556-4a55-86fc-9c97363cc24d", "address": "fa:16:3e:8b:c0:4d", "network": {"id": "adf4395d-70f5-4db9-9730-2fd88ca189a9", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-307667077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e1187e438e194c4587f8824110dc23f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap299ec13e-c5", "ovs_interfaceid": "299ec13e-c556-4a55-86fc-9c97363cc24d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 897.577204] env[65385]: DEBUG nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 897.623475] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd6ff46-a5b9-4b6e-86c2-14803b9d1e7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.628862] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.635459] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fb8448-884a-4b3d-aa55-1dd7a5deef2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.668739] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73677050-d2e8-4fae-bf72-156364e06558 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.677197] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc63ba9a-3d05-4be4-ae24-80ae88a1af22 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.692520] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Releasing lock "refresh_cache-cf37e10a-14f4-4456-bad0-d7528457658d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 897.693903] env[65385]: DEBUG nova.compute.provider_tree [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.732373] env[65385]: DEBUG nova.compute.manager [req-678cde0d-5fd1-4422-971f-73c24bdd5e0c req-693fa0fb-78de-47e5-ae90-aff66f1d8fa3 service nova] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Received event network-vif-deleted-15a6f057-527a-433e-bb76-8d2b57f48765 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 897.821576] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.821576] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95646476-d809-4204-a74f-ad13c907288d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.828451] env[65385]: DEBUG oslo_vmware.api [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 897.828451] env[65385]: value = "task-4453932" [ 897.828451] env[65385]: _type = "Task" [ 897.828451] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.838350] env[65385]: DEBUG oslo_vmware.api [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453932, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.109051] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.200066] env[65385]: DEBUG nova.scheduler.client.report [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 898.204596] env[65385]: DEBUG nova.compute.manager [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 898.205830] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aff3fde-58c5-418f-89e1-4f55d0ff8e54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.339027] env[65385]: DEBUG oslo_vmware.api [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453932, 'name': PowerOnVM_Task, 'duration_secs': 0.408448} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.339027] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.339236] env[65385]: DEBUG nova.compute.manager [None req-dd08ac8a-69fa-402d-a63c-2fd58d04904f tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 898.340034] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601e14b0-07da-4212-a7cf-20af172c9b09 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.713527] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.714137] env[65385]: DEBUG nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 898.717429] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.631s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.719151] env[65385]: INFO nova.compute.claims [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.225490] env[65385]: DEBUG nova.compute.utils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 899.229773] env[65385]: DEBUG nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 899.230028] env[65385]: DEBUG nova.network.neutron [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 899.230335] env[65385]: WARNING neutronclient.v2_0.client [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.230635] env[65385]: WARNING neutronclient.v2_0.client [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.231344] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.231712] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.240576] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcc34e6-1af8-4a9e-a1ee-c7be5189de44 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.249512] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Doing hard reboot of VM {{(pid=65385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 899.249849] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-c743611e-872c-4ec9-b796-12331b3c1c1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.258143] env[65385]: DEBUG oslo_vmware.api [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 899.258143] env[65385]: value = "task-4453933" [ 899.258143] env[65385]: _type = "Task" [ 899.258143] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.267015] env[65385]: DEBUG oslo_vmware.api [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453933, 'name': ResetVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.315277] env[65385]: DEBUG nova.policy [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03f7c78b16e547c38f24ed2c8a329202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ba61d77cc2d4fe1ba3c03466d5985d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 899.731057] env[65385]: DEBUG nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 899.758618] env[65385]: DEBUG nova.network.neutron [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Successfully created port: 2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 899.773936] env[65385]: DEBUG oslo_vmware.api [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453933, 'name': ResetVM_Task, 'duration_secs': 0.116621} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.774231] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Did hard reboot of VM {{(pid=65385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 899.774509] env[65385]: DEBUG nova.compute.manager [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 899.775497] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e3bcba-861e-4b3c-8946-9c46b450fa10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.166979] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72552239-ee16-4456-b643-db30e6f700a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.175238] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516406f6-8e1a-45a1-ae2a-cd6aac3ab730 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.210123] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da775a06-6ade-47a7-b2ff-8b4a15c9c45d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.219036] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd74d5c3-25be-43f0-9804-282258ba3b81 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.235218] env[65385]: DEBUG nova.compute.provider_tree [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.296698] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d6fa92d0-528e-4fcc-932b-0583724886cc tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.542s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.509033] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.509491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.509627] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.509874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.511235] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.513694] env[65385]: INFO nova.compute.manager [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Terminating instance [ 900.742923] env[65385]: DEBUG nova.scheduler.client.report [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 900.750841] env[65385]: DEBUG nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 900.791292] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 900.791556] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 900.791707] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 900.791881] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 900.792031] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 900.792181] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 900.792397] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.792552] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 900.792713] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 900.792870] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 900.793058] env[65385]: DEBUG nova.virt.hardware [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 900.794068] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5391f4-b577-4e84-a6ee-40365455ae0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.807242] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355be1f5-6679-46df-9a52-1cf4939b8aff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.018860] env[65385]: DEBUG nova.compute.manager [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 901.019133] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.021844] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b2aac3-a768-44f3-a556-bd328b8f4721 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.029948] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.031315] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49046152-755c-4270-9cb9-7e002bb1158e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.044287] env[65385]: DEBUG oslo_vmware.api [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 901.044287] env[65385]: value = "task-4453934" [ 901.044287] env[65385]: _type = "Task" [ 901.044287] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.055242] env[65385]: DEBUG oslo_vmware.api [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.254301] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.254628] env[65385]: DEBUG nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 901.262037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.020s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.262037] env[65385]: DEBUG nova.objects.instance [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lazy-loading 'resources' on Instance uuid 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.319970] env[65385]: DEBUG nova.compute.manager [req-ae238962-2921-4a16-a12a-668f549a2371 req-7f01b7e6-c7bf-4a15-af45-1bf2eb0af220 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Received event network-vif-plugged-2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 901.320264] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae238962-2921-4a16-a12a-668f549a2371 req-7f01b7e6-c7bf-4a15-af45-1bf2eb0af220 service nova] Acquiring lock "4b0b8859-b63e-4740-bf67-5733c75e1c70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.320453] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae238962-2921-4a16-a12a-668f549a2371 req-7f01b7e6-c7bf-4a15-af45-1bf2eb0af220 service nova] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.320806] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae238962-2921-4a16-a12a-668f549a2371 req-7f01b7e6-c7bf-4a15-af45-1bf2eb0af220 service nova] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.320806] env[65385]: DEBUG nova.compute.manager [req-ae238962-2921-4a16-a12a-668f549a2371 req-7f01b7e6-c7bf-4a15-af45-1bf2eb0af220 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] No waiting events found dispatching network-vif-plugged-2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 901.320957] env[65385]: WARNING nova.compute.manager [req-ae238962-2921-4a16-a12a-668f549a2371 req-7f01b7e6-c7bf-4a15-af45-1bf2eb0af220 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Received unexpected event network-vif-plugged-2bb8c7e3-459c-4149-8960-0d0df350dc4a for instance with vm_state building and task_state spawning. [ 901.365056] env[65385]: DEBUG nova.network.neutron [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Successfully updated port: 2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 901.538528] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "cf37e10a-14f4-4456-bad0-d7528457658d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.538791] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.539848] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "cf37e10a-14f4-4456-bad0-d7528457658d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.541218] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.541456] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.544881] env[65385]: INFO nova.compute.manager [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Terminating instance [ 901.564969] env[65385]: DEBUG oslo_vmware.api [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453934, 'name': PowerOffVM_Task, 'duration_secs': 0.20639} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.566029] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.566228] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.567408] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c05ea213-00da-4308-bab4-c9d125450cd8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.641591] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.641591] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.641591] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleting the datastore file [datastore2] 89f3c815-8671-47ce-9e74-bf6e652bb3c3 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.641591] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-902eebbf-1982-4548-bb0a-8a8f37b00bbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.648190] env[65385]: DEBUG oslo_vmware.api [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 901.648190] env[65385]: value = "task-4453936" [ 901.648190] env[65385]: _type = "Task" [ 901.648190] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.661591] env[65385]: DEBUG oslo_vmware.api [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.762761] env[65385]: DEBUG nova.compute.utils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 901.766748] env[65385]: DEBUG nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 901.766969] env[65385]: DEBUG nova.network.neutron [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 901.768624] env[65385]: WARNING neutronclient.v2_0.client [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.768624] env[65385]: WARNING neutronclient.v2_0.client [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.768624] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.768725] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.829446] env[65385]: DEBUG nova.policy [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77dc7acbf0324ac88b7092803faaa384', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b27c092a0fd345ea953956469c21f8bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 901.874898] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.874898] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.874898] env[65385]: DEBUG nova.network.neutron [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 902.057314] env[65385]: DEBUG nova.compute.manager [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 902.057473] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.058368] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109a9133-39e9-4430-8bab-93c1629cf133 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.067755] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.071307] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d8a772d-bb2c-4962-b2d8-ccd1fcde2cb4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.079581] env[65385]: DEBUG oslo_vmware.api [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 902.079581] env[65385]: value = "task-4453937" [ 902.079581] env[65385]: _type = "Task" [ 902.079581] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.091576] env[65385]: DEBUG oslo_vmware.api [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.160231] env[65385]: DEBUG oslo_vmware.api [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4453936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150107} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.160231] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.160231] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 902.160480] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 902.160480] env[65385]: INFO nova.compute.manager [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Took 1.14 seconds to destroy the instance on the hypervisor. [ 902.160639] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 902.163030] env[65385]: DEBUG nova.compute.manager [-] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 902.163122] env[65385]: DEBUG nova.network.neutron [-] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 902.163778] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 902.164339] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.164653] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.171942] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10065c71-5732-42e3-89be-49df31ca942d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.181715] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8eaf47-73e7-46d0-a3be-9be96cf92f50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.212966] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 902.215436] env[65385]: DEBUG nova.network.neutron [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Successfully created port: f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 902.218248] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6456f482-d107-4f09-a59d-34e5a48a9ee5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.228580] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a16fc92-322e-43b2-9044-78ed496d0d74 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.243547] env[65385]: DEBUG nova.compute.provider_tree [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.277165] env[65385]: DEBUG nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 902.377614] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.379189] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.591257] env[65385]: DEBUG oslo_vmware.api [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453937, 'name': PowerOffVM_Task, 'duration_secs': 0.195503} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.591472] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.591634] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.591890] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b72b5dc-5c38-4447-92a3-3f4d4813e60b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.665029] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.665029] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.665499] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Deleting the datastore file [datastore2] cf37e10a-14f4-4456-bad0-d7528457658d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.667797] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9052c9d-d338-405a-a43b-c445a3d0009b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.676124] env[65385]: DEBUG oslo_vmware.api [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for the task: (returnval){ [ 902.676124] env[65385]: value = "task-4453939" [ 902.676124] env[65385]: _type = "Task" [ 902.676124] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.685439] env[65385]: DEBUG oslo_vmware.api [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.739769] env[65385]: DEBUG nova.network.neutron [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 902.747140] env[65385]: DEBUG nova.scheduler.client.report [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 902.781533] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.781955] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.930932] env[65385]: WARNING neutronclient.v2_0.client [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 902.931832] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.932753] env[65385]: WARNING openstack [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.047363] env[65385]: DEBUG nova.compute.manager [req-95ce8af1-c93e-4966-9fa2-b94292e5d513 req-45111127-6b4b-41e6-973a-55301dd06146 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Received event network-vif-deleted-18c18d61-0493-40e3-8883-c90faabf147b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 903.047710] env[65385]: INFO nova.compute.manager [req-95ce8af1-c93e-4966-9fa2-b94292e5d513 req-45111127-6b4b-41e6-973a-55301dd06146 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Neutron deleted interface 18c18d61-0493-40e3-8883-c90faabf147b; detaching it from the instance and deleting it from the info cache [ 903.047878] env[65385]: DEBUG nova.network.neutron [req-95ce8af1-c93e-4966-9fa2-b94292e5d513 req-45111127-6b4b-41e6-973a-55301dd06146 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.055570] env[65385]: DEBUG nova.network.neutron [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Updating instance_info_cache with network_info: [{"id": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "address": "fa:16:3e:b8:39:d3", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bb8c7e3-45", "ovs_interfaceid": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.172155] env[65385]: DEBUG nova.network.neutron [-] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.188615] env[65385]: DEBUG oslo_vmware.api [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Task: {'id': task-4453939, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129675} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.188903] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.189097] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.189274] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.189438] env[65385]: INFO nova.compute.manager [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 903.189697] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 903.189861] env[65385]: DEBUG nova.compute.manager [-] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 903.189960] env[65385]: DEBUG nova.network.neutron [-] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 903.190384] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 903.190927] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.191308] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.236914] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 903.255135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.996s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.258684] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.170s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.260219] env[65385]: INFO nova.compute.claims [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.290566] env[65385]: INFO nova.scheduler.client.report [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Deleted allocations for instance 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0 [ 903.299412] env[65385]: DEBUG nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 903.336133] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 903.336444] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 903.336889] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 903.336889] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 903.337048] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 903.337151] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 903.337361] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.337536] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 903.337673] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 903.338118] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 903.338118] env[65385]: DEBUG nova.virt.hardware [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 903.340350] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c355585-622e-4670-95d9-7e0f0bf53c8e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.350267] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a80132-f192-448c-86ef-9206eba72c79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.559300] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.559673] env[65385]: DEBUG nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Instance network_info: |[{"id": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "address": "fa:16:3e:b8:39:d3", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bb8c7e3-45", "ovs_interfaceid": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 903.560672] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:39:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2bb8c7e3-459c-4149-8960-0d0df350dc4a', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.568258] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 903.568498] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9f2570f-efae-4a86-a523-009950d9835d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.570603] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.570825] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f77ef151-bb66-4bbb-9efe-c9433468e508 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.597024] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03354b2e-e182-4bad-80d3-96aea1c296f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.609866] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.609866] env[65385]: value = "task-4453940" [ 903.609866] env[65385]: _type = "Task" [ 903.609866] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.630690] env[65385]: DEBUG nova.compute.manager [req-95ce8af1-c93e-4966-9fa2-b94292e5d513 req-45111127-6b4b-41e6-973a-55301dd06146 service nova] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Detach interface failed, port_id=18c18d61-0493-40e3-8883-c90faabf147b, reason: Instance 89f3c815-8671-47ce-9e74-bf6e652bb3c3 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 903.634061] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453940, 'name': CreateVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.675098] env[65385]: INFO nova.compute.manager [-] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Took 1.51 seconds to deallocate network for instance. [ 903.718497] env[65385]: DEBUG nova.compute.manager [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Received event network-changed-2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 903.718686] env[65385]: DEBUG nova.compute.manager [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Refreshing instance network info cache due to event network-changed-2bb8c7e3-459c-4149-8960-0d0df350dc4a. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 903.718897] env[65385]: DEBUG oslo_concurrency.lockutils [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Acquiring lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.719080] env[65385]: DEBUG oslo_concurrency.lockutils [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Acquired lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.719242] env[65385]: DEBUG nova.network.neutron [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Refreshing network info cache for port 2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 903.800284] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6fa517df-d5a6-4d5a-94b1-b3637045faa0 tempest-ServersWithSpecificFlavorTestJSON-1346103607 tempest-ServersWithSpecificFlavorTestJSON-1346103607-project-member] Lock "6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.693s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.813197] env[65385]: DEBUG nova.network.neutron [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Successfully updated port: f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 903.962433] env[65385]: DEBUG nova.network.neutron [-] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 904.124192] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453940, 'name': CreateVM_Task, 'duration_secs': 0.381558} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.125362] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.125894] env[65385]: WARNING neutronclient.v2_0.client [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.126349] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.126523] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.126842] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 904.127172] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-903084d6-58d0-4214-b273-eeb05735553e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.138495] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 904.138495] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525fd934-8ed3-ccc0-dd9e-c2d2a99ee00a" [ 904.138495] env[65385]: _type = "Task" [ 904.138495] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.149811] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525fd934-8ed3-ccc0-dd9e-c2d2a99ee00a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.181869] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.222724] env[65385]: WARNING neutronclient.v2_0.client [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.223441] env[65385]: WARNING openstack [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.223797] env[65385]: WARNING openstack [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.316620] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.316958] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.317185] env[65385]: DEBUG nova.network.neutron [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 904.334028] env[65385]: WARNING openstack [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.334505] env[65385]: WARNING openstack [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.419615] env[65385]: WARNING neutronclient.v2_0.client [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.420132] env[65385]: WARNING openstack [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.420874] env[65385]: WARNING openstack [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.466008] env[65385]: INFO nova.compute.manager [-] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Took 1.28 seconds to deallocate network for instance. [ 904.560753] env[65385]: DEBUG nova.network.neutron [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Updated VIF entry in instance network info cache for port 2bb8c7e3-459c-4149-8960-0d0df350dc4a. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 904.560984] env[65385]: DEBUG nova.network.neutron [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Updating instance_info_cache with network_info: [{"id": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "address": "fa:16:3e:b8:39:d3", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bb8c7e3-45", "ovs_interfaceid": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 904.626987] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5383972f-032e-4626-a7a8-b5e0e75cc00e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.635826] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2b9ffa-7ae3-4706-9233-2db29d2e8c21 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.650699] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525fd934-8ed3-ccc0-dd9e-c2d2a99ee00a, 'name': SearchDatastore_Task, 'duration_secs': 0.011442} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.682108] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 904.682108] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.682108] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.682108] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.682108] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.682108] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2ca6388-f093-4eaf-9691-725ba1d0cadd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.684568] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057db9a5-0ffd-4ebe-81a6-b09408b8a723 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.692898] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2dd37a-9f35-45d7-8a18-c62ee1a6d51e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.697739] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.697922] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 904.698980] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f6130ad-9ef0-4f71-9b2f-77357aeba829 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.709728] env[65385]: DEBUG nova.compute.provider_tree [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.716316] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 904.716316] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e5b9e-cb47-56a8-7418-773855dbdb54" [ 904.716316] env[65385]: _type = "Task" [ 904.716316] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.724442] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e5b9e-cb47-56a8-7418-773855dbdb54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.821411] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.821574] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.862930] env[65385]: DEBUG nova.network.neutron [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 904.883970] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.884394] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.933016] env[65385]: WARNING neutronclient.v2_0.client [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.933812] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.934246] env[65385]: WARNING openstack [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.979235] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.045218] env[65385]: DEBUG nova.network.neutron [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 905.064199] env[65385]: DEBUG oslo_concurrency.lockutils [req-53f351b8-ce61-49e8-9268-cbda0a715652 req-a26411f6-0b97-42c2-945b-f0c180fbb492 service nova] Releasing lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.134937] env[65385]: DEBUG nova.compute.manager [req-9401346e-cf8b-46f0-819b-e7d8fa86f3c8 req-483997a2-6a3d-4cc7-bf49-e01802e1aa3b service nova] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Received event network-vif-deleted-299ec13e-c556-4a55-86fc-9c97363cc24d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 905.214066] env[65385]: DEBUG nova.scheduler.client.report [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.230498] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e5b9e-cb47-56a8-7418-773855dbdb54, 'name': SearchDatastore_Task, 'duration_secs': 0.019125} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.231459] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd9347bc-2ead-4794-bb20-c556c0b0cac1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.238047] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 905.238047] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52356eb3-9ddc-9e92-15ad-f279741a6197" [ 905.238047] env[65385]: _type = "Task" [ 905.238047] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.250743] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52356eb3-9ddc-9e92-15ad-f279741a6197, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.549832] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.553017] env[65385]: DEBUG nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Instance network_info: |[{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 905.553017] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:13:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4a8fd90-153b-494f-b76a-299eb05c03f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7b3d6ae-eb2d-43c1-8c44-ca85849be867', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.565547] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating folder: Project (b27c092a0fd345ea953956469c21f8bb). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.565547] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc64e506-b856-430c-9af3-28b5bd66df8a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.575933] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created folder: Project (b27c092a0fd345ea953956469c21f8bb) in parent group-v870881. [ 905.576326] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating folder: Instances. Parent ref: group-v871090. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.576618] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d83ab9ce-e130-490d-b57b-af5690bbb236 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.588395] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created folder: Instances in parent group-v871090. [ 905.589100] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 905.589339] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.589616] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6c437ece-ebe3-4a9b-a153-935edcc9b713 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.616702] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.616702] env[65385]: value = "task-4453943" [ 905.616702] env[65385]: _type = "Task" [ 905.616702] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.626753] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453943, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.728039] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.728039] env[65385]: DEBUG nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 905.730517] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.887s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.730870] env[65385]: DEBUG nova.objects.instance [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lazy-loading 'resources' on Instance uuid 0b02fcd5-9fc7-4543-a754-d31fa0186981 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.754136] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52356eb3-9ddc-9e92-15ad-f279741a6197, 'name': SearchDatastore_Task, 'duration_secs': 0.017781} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.754136] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.754136] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 4b0b8859-b63e-4740-bf67-5733c75e1c70/4b0b8859-b63e-4740-bf67-5733c75e1c70.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.754136] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d68ad58-3324-4dac-a403-73235fc108c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.762133] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 905.762133] env[65385]: value = "task-4453944" [ 905.762133] env[65385]: _type = "Task" [ 905.762133] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.780628] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.805657] env[65385]: DEBUG nova.compute.manager [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-vif-plugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 905.806800] env[65385]: DEBUG oslo_concurrency.lockutils [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.807081] env[65385]: DEBUG oslo_concurrency.lockutils [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.812200] env[65385]: DEBUG oslo_concurrency.lockutils [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.812200] env[65385]: DEBUG nova.compute.manager [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] No waiting events found dispatching network-vif-plugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 905.812200] env[65385]: WARNING nova.compute.manager [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received unexpected event network-vif-plugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 for instance with vm_state building and task_state spawning. [ 905.812200] env[65385]: DEBUG nova.compute.manager [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 905.812200] env[65385]: DEBUG nova.compute.manager [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing instance network info cache due to event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 905.812200] env[65385]: DEBUG oslo_concurrency.lockutils [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Acquiring lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.812200] env[65385]: DEBUG oslo_concurrency.lockutils [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Acquired lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.812200] env[65385]: DEBUG nova.network.neutron [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 906.128505] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453943, 'name': CreateVM_Task, 'duration_secs': 0.354532} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.128793] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.129287] env[65385]: WARNING neutronclient.v2_0.client [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.129720] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.129883] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.130316] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 906.130624] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b2db52a-e345-4872-8792-e2d150f93ad7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.138248] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 906.138248] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5220939b-9e92-d77a-c538-5d35943141b4" [ 906.138248] env[65385]: _type = "Task" [ 906.138248] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.148233] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5220939b-9e92-d77a-c538-5d35943141b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.242108] env[65385]: DEBUG nova.compute.utils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 906.243925] env[65385]: DEBUG nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 906.248163] env[65385]: DEBUG nova.network.neutron [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 906.248637] env[65385]: WARNING neutronclient.v2_0.client [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.248885] env[65385]: WARNING neutronclient.v2_0.client [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.249539] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.249924] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.260756] env[65385]: DEBUG nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 906.283134] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479537} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.283244] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 4b0b8859-b63e-4740-bf67-5733c75e1c70/4b0b8859-b63e-4740-bf67-5733c75e1c70.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.284902] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.284902] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be56799b-a817-43f4-83f7-cff0de9d75da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.297156] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 906.297156] env[65385]: value = "task-4453945" [ 906.297156] env[65385]: _type = "Task" [ 906.297156] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.306699] env[65385]: DEBUG nova.policy [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0f069272314a31ad33c77137513a17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c27eb6232a2148a1a259f57494b4ae30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 906.314683] env[65385]: WARNING neutronclient.v2_0.client [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.315667] env[65385]: WARNING openstack [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.315916] env[65385]: WARNING openstack [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.333172] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453945, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.655233] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5220939b-9e92-d77a-c538-5d35943141b4, 'name': SearchDatastore_Task, 'duration_secs': 0.037763} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.659232] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.659965] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.660272] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.660431] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.660610] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.661084] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e49914d-4170-4062-b4b9-4f7fad95dc50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.671623] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.671623] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.672338] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8974e2c7-de37-43f5-9571-835d7cf70e19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.681398] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 906.681398] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5275922d-a13d-0f66-e8c2-d32701f812e3" [ 906.681398] env[65385]: _type = "Task" [ 906.681398] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.690921] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5275922d-a13d-0f66-e8c2-d32701f812e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.692639] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-143005b2-5539-4390-83c3-94c876d351af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.700049] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bee8e04-af85-4d81-bb97-2b1383a5cedf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.736145] env[65385]: DEBUG nova.network.neutron [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Successfully created port: 6785f095-0836-4bc7-923a-b16075641a80 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 906.739689] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3f6e74-0f06-44df-98b7-be4d784b8074 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.747743] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53823b6-fa13-43b3-87d1-91cc7ad1dbf8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.768836] env[65385]: DEBUG nova.compute.provider_tree [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.806469] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453945, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078592} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.806789] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.807660] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ebf954-1c79-45d2-817f-1522acae21f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.840746] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 4b0b8859-b63e-4740-bf67-5733c75e1c70/4b0b8859-b63e-4740-bf67-5733c75e1c70.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.847027] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6b425ce-8b99-457e-90a3-6a1619baacb1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.869612] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 906.869612] env[65385]: value = "task-4453946" [ 906.869612] env[65385]: _type = "Task" [ 906.869612] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.879384] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.096481] env[65385]: WARNING openstack [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.096867] env[65385]: WARNING openstack [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.192498] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5275922d-a13d-0f66-e8c2-d32701f812e3, 'name': SearchDatastore_Task, 'duration_secs': 0.010633} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.193301] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8caaff41-cd46-45d8-9487-64bf1ff34dba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.198693] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 907.198693] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa05d2-9802-4e49-0415-16b3f4211e9d" [ 907.198693] env[65385]: _type = "Task" [ 907.198693] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.207012] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa05d2-9802-4e49-0415-16b3f4211e9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.271694] env[65385]: DEBUG nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 907.274573] env[65385]: DEBUG nova.scheduler.client.report [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.298881] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='eee8cf5e4961bdb48c4fad05c8a9e747',container_format='bare',created_at=2025-11-14T16:51:43Z,direct_url=,disk_format='vmdk',id=57fab5fe-ac7e-4d91-94ca-0285c2957e33,min_disk=1,min_ram=0,name='tempest-test-snap-971896988',owner='c27eb6232a2148a1a259f57494b4ae30',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-11-14T16:51:56Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 907.299147] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 907.299301] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 907.299484] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 907.299629] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 907.299771] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 907.300020] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.300312] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 907.300487] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 907.300649] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 907.300820] env[65385]: DEBUG nova.virt.hardware [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 907.301988] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fb4bea-dae9-4714-b5a6-d7ad167b02f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.311744] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2d5769-0025-465e-b156-f265574e442e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.381051] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453946, 'name': ReconfigVM_Task, 'duration_secs': 0.282398} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.381345] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 4b0b8859-b63e-4740-bf67-5733c75e1c70/4b0b8859-b63e-4740-bf67-5733c75e1c70.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.381986] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69fd6c81-5d99-4cac-8a81-8fd68a3f9f9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.390422] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 907.390422] env[65385]: value = "task-4453947" [ 907.390422] env[65385]: _type = "Task" [ 907.390422] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.399347] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453947, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.710750] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa05d2-9802-4e49-0415-16b3f4211e9d, 'name': SearchDatastore_Task, 'duration_secs': 0.010287} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.711131] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.711333] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.711643] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6432caf5-754f-455d-9da5-2ef2bb51c3aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.720019] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 907.720019] env[65385]: value = "task-4453948" [ 907.720019] env[65385]: _type = "Task" [ 907.720019] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.729923] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453948, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.779197] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.048s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.782720] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.683s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 907.783763] env[65385]: INFO nova.compute.claims [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.823863] env[65385]: INFO nova.scheduler.client.report [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Deleted allocations for instance 0b02fcd5-9fc7-4543-a754-d31fa0186981 [ 907.907401] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453947, 'name': Rename_Task, 'duration_secs': 0.160288} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.907753] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.908160] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fe3fff5-eeac-42b4-a42a-93c7f6ceb400 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.919075] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 907.919075] env[65385]: value = "task-4453949" [ 907.919075] env[65385]: _type = "Task" [ 907.919075] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.936108] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453949, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.235896] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453948, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.270470] env[65385]: WARNING neutronclient.v2_0.client [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.271166] env[65385]: WARNING openstack [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 908.271544] env[65385]: WARNING openstack [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 908.319187] env[65385]: DEBUG nova.network.neutron [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Successfully updated port: 6785f095-0836-4bc7-923a-b16075641a80 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 908.336794] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdaed69c-a25b-47bb-9c24-e6736ae0c41b tempest-ServerAddressesNegativeTestJSON-1213233335 tempest-ServerAddressesNegativeTestJSON-1213233335-project-member] Lock "0b02fcd5-9fc7-4543-a754-d31fa0186981" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.347s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.420423] env[65385]: DEBUG nova.network.neutron [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updated VIF entry in instance network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 908.420868] env[65385]: DEBUG nova.network.neutron [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 908.442458] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453949, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.732953] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453948, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530176} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.733897] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.733897] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.733897] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0966ba9-760b-46e8-b01e-29be26555bb6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.742778] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 908.742778] env[65385]: value = "task-4453950" [ 908.742778] env[65385]: _type = "Task" [ 908.742778] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.754051] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453950, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.826128] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "refresh_cache-ac6fa516-d954-466e-b8e0-e12440492049" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.826479] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "refresh_cache-ac6fa516-d954-466e-b8e0-e12440492049" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.826758] env[65385]: DEBUG nova.network.neutron [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 908.927363] env[65385]: DEBUG oslo_concurrency.lockutils [req-7a95cca9-9b11-4756-86e4-f0a11492c455 req-1770cf72-e6da-4728-bd81-846a81d910dc service nova] Releasing lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.938556] env[65385]: DEBUG oslo_vmware.api [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4453949, 'name': PowerOnVM_Task, 'duration_secs': 0.643978} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.938872] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.939860] env[65385]: INFO nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Took 8.19 seconds to spawn the instance on the hypervisor. [ 908.939860] env[65385]: DEBUG nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 908.943154] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663941d7-ed25-4721-bed1-c23cd98c3c23 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.029844] env[65385]: DEBUG nova.compute.manager [req-ac5b6022-d71c-4902-ad15-ef341a9a1646 req-da338c84-498a-46f9-9203-902a1ccdc5d7 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Received event network-vif-plugged-6785f095-0836-4bc7-923a-b16075641a80 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 909.030024] env[65385]: DEBUG oslo_concurrency.lockutils [req-ac5b6022-d71c-4902-ad15-ef341a9a1646 req-da338c84-498a-46f9-9203-902a1ccdc5d7 service nova] Acquiring lock "ac6fa516-d954-466e-b8e0-e12440492049-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.030233] env[65385]: DEBUG oslo_concurrency.lockutils [req-ac5b6022-d71c-4902-ad15-ef341a9a1646 req-da338c84-498a-46f9-9203-902a1ccdc5d7 service nova] Lock "ac6fa516-d954-466e-b8e0-e12440492049-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.030566] env[65385]: DEBUG oslo_concurrency.lockutils [req-ac5b6022-d71c-4902-ad15-ef341a9a1646 req-da338c84-498a-46f9-9203-902a1ccdc5d7 service nova] Lock "ac6fa516-d954-466e-b8e0-e12440492049-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.030626] env[65385]: DEBUG nova.compute.manager [req-ac5b6022-d71c-4902-ad15-ef341a9a1646 req-da338c84-498a-46f9-9203-902a1ccdc5d7 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] No waiting events found dispatching network-vif-plugged-6785f095-0836-4bc7-923a-b16075641a80 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 909.030752] env[65385]: WARNING nova.compute.manager [req-ac5b6022-d71c-4902-ad15-ef341a9a1646 req-da338c84-498a-46f9-9203-902a1ccdc5d7 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Received unexpected event network-vif-plugged-6785f095-0836-4bc7-923a-b16075641a80 for instance with vm_state building and task_state spawning. [ 909.174702] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25460ce-2679-4008-bc49-33c9a4d1fbf0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.185503] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f4e958-46cb-4213-8c67-f5f3b239b7a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.222418] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a8592f-7ba0-4898-a33d-1e0c65ec1230 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.231015] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a45258-03d3-4c14-821a-bc59bdbd06be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.246051] env[65385]: DEBUG nova.compute.provider_tree [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.256580] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453950, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.218124} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.257729] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.258628] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623e652b-b2be-4b03-a767-b43fea84c697 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.284337] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.285060] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8db0d3d-099e-443e-9caf-d73d98abbe58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.308855] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 909.308855] env[65385]: value = "task-4453951" [ 909.308855] env[65385]: _type = "Task" [ 909.308855] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.324479] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.332229] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.332786] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.442905] env[65385]: DEBUG nova.network.neutron [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 909.463927] env[65385]: INFO nova.compute.manager [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Took 36.19 seconds to build instance. [ 909.626341] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.626786] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.750397] env[65385]: WARNING neutronclient.v2_0.client [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 909.751819] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.751819] env[65385]: WARNING openstack [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.762083] env[65385]: DEBUG nova.scheduler.client.report [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 909.823480] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453951, 'name': ReconfigVM_Task, 'duration_secs': 0.299713} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.823961] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.824603] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90e71b6e-f37d-4520-9b8a-1e7b781b3b10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.832442] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 909.832442] env[65385]: value = "task-4453952" [ 909.832442] env[65385]: _type = "Task" [ 909.832442] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.842979] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453952, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.930571] env[65385]: DEBUG nova.network.neutron [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Updating instance_info_cache with network_info: [{"id": "6785f095-0836-4bc7-923a-b16075641a80", "address": "fa:16:3e:42:7c:73", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6785f095-08", "ovs_interfaceid": "6785f095-0836-4bc7-923a-b16075641a80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 909.966430] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb8c180e-6197-4492-9b12-17efe52aa963 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.706s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.267607] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.268167] env[65385]: DEBUG nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 910.273016] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.296s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.273016] env[65385]: DEBUG nova.objects.instance [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lazy-loading 'resources' on Instance uuid af8d01e2-fb86-41c4-99a5-204b30eeda0c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.348934] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453952, 'name': Rename_Task, 'duration_secs': 0.221288} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.349242] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.349505] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f700ddad-f87e-4706-a434-8abd6813b078 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.358434] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 910.358434] env[65385]: value = "task-4453953" [ 910.358434] env[65385]: _type = "Task" [ 910.358434] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.372914] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453953, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.435628] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "refresh_cache-ac6fa516-d954-466e-b8e0-e12440492049" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 910.436055] env[65385]: DEBUG nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Instance network_info: |[{"id": "6785f095-0836-4bc7-923a-b16075641a80", "address": "fa:16:3e:42:7c:73", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6785f095-08", "ovs_interfaceid": "6785f095-0836-4bc7-923a-b16075641a80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 910.436575] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:7c:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6785f095-0836-4bc7-923a-b16075641a80', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.448361] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 910.448361] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.448361] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8955e532-af1d-4e42-becf-e8e29a380202 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.470449] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.470449] env[65385]: value = "task-4453954" [ 910.470449] env[65385]: _type = "Task" [ 910.470449] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.480936] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453954, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.776272] env[65385]: DEBUG nova.compute.utils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 910.777061] env[65385]: DEBUG nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 910.777253] env[65385]: DEBUG nova.network.neutron [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 910.777601] env[65385]: WARNING neutronclient.v2_0.client [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.777931] env[65385]: WARNING neutronclient.v2_0.client [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.778533] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.778872] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.875907] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453953, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.904505] env[65385]: DEBUG nova.policy [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d5103a5769a4296973351194f49fed3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '754b1facaaa14501b2204c98e1d7a5a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 910.987198] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453954, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.157273] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79a8aa4-bc3c-42f2-8ec9-47c74a131957 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.170289] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e9ed98-81aa-4d94-ae44-ad47c42d0962 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.221797] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbd23ca-8d21-44f2-b769-24530efffbb1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.230718] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4f10ea-db54-40cd-8c44-8240ee92d2e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.249796] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.251650] env[65385]: DEBUG nova.compute.provider_tree [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.254468] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.290205] env[65385]: DEBUG nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 911.371591] env[65385]: DEBUG oslo_vmware.api [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4453953, 'name': PowerOnVM_Task, 'duration_secs': 0.60716} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.371882] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.372164] env[65385]: INFO nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Took 8.07 seconds to spawn the instance on the hypervisor. [ 911.372821] env[65385]: DEBUG nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 911.373243] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2377b7ab-2936-4e90-b0be-d68fa230984a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.486036] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453954, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.524766] env[65385]: DEBUG nova.network.neutron [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Successfully created port: 1a20c5c9-1b05-42cf-a71f-894e2921e893 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 911.759498] env[65385]: DEBUG nova.scheduler.client.report [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 911.772434] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.773172] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.773172] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.773172] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.773172] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.773641] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.773641] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 911.773641] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.894895] env[65385]: INFO nova.compute.manager [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Took 31.83 seconds to build instance. [ 911.985965] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453954, 'name': CreateVM_Task, 'duration_secs': 1.358876} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.986160] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.986676] env[65385]: WARNING neutronclient.v2_0.client [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.987126] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.987289] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.987651] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 911.987909] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40d3793c-0c33-42f1-9ad5-890c45f4caac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.994256] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 911.994256] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fd0b4d-0a37-007e-68cf-75f128fff24c" [ 911.994256] env[65385]: _type = "Task" [ 911.994256] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.003644] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fd0b4d-0a37-007e-68cf-75f128fff24c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.273020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.274644] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.287s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.275836] env[65385]: DEBUG nova.objects.instance [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lazy-loading 'resources' on Instance uuid 42b9edf6-1873-49c3-8074-8eef654ac371 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.277892] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 912.304178] env[65385]: DEBUG nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 912.313940] env[65385]: INFO nova.scheduler.client.report [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Deleted allocations for instance af8d01e2-fb86-41c4-99a5-204b30eeda0c [ 912.347946] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 912.348228] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 912.348405] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 912.348577] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 912.348762] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 912.348932] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 912.349162] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.349318] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 912.349480] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 912.349638] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 912.349883] env[65385]: DEBUG nova.virt.hardware [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 912.350870] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d95cc9-a05a-4d88-8639-599dc24a7254 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.360516] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218d1bdf-4aa3-4ed7-9149-731feb18f978 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.398043] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3ce65f86-7942-4004-b579-f61a70ab0f02 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.861s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.514614] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 912.514614] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Processing image 57fab5fe-ac7e-4d91-94ca-0285c2957e33 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.514816] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33/57fab5fe-ac7e-4d91-94ca-0285c2957e33.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.517014] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33/57fab5fe-ac7e-4d91-94ca-0285c2957e33.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.517014] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.517014] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d3c7647-3764-47a5-a4e5-1432dac476e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.528040] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.528040] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.528223] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97dea6a0-a611-48db-81ae-4f73b1612011 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.536318] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 912.536318] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fc73bd-645d-2cc6-12fb-d61d572d9fa5" [ 912.536318] env[65385]: _type = "Task" [ 912.536318] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.546127] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fc73bd-645d-2cc6-12fb-d61d572d9fa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.823742] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfea82db-3fc9-4409-adcc-1b191b552d93 tempest-ImagesOneServerTestJSON-386490806 tempest-ImagesOneServerTestJSON-386490806-project-member] Lock "af8d01e2-fb86-41c4-99a5-204b30eeda0c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.286s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.056951] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Preparing fetch location {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 913.057573] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Fetch image to [datastore2] OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3/OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3.vmdk {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 913.058392] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Downloading stream optimized image 57fab5fe-ac7e-4d91-94ca-0285c2957e33 to [datastore2] OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3/OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3.vmdk on the data store datastore2 as vApp {{(pid=65385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 913.058731] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Downloading image file data 57fab5fe-ac7e-4d91-94ca-0285c2957e33 to the ESX as VM named 'OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3' {{(pid=65385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 913.134672] env[65385]: DEBUG nova.compute.manager [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Received event network-changed-6785f095-0836-4bc7-923a-b16075641a80 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 913.134918] env[65385]: DEBUG nova.compute.manager [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Refreshing instance network info cache due to event network-changed-6785f095-0836-4bc7-923a-b16075641a80. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 913.135365] env[65385]: DEBUG oslo_concurrency.lockutils [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Acquiring lock "refresh_cache-ac6fa516-d954-466e-b8e0-e12440492049" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.135365] env[65385]: DEBUG oslo_concurrency.lockutils [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Acquired lock "refresh_cache-ac6fa516-d954-466e-b8e0-e12440492049" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 913.135365] env[65385]: DEBUG nova.network.neutron [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Refreshing network info cache for port 6785f095-0836-4bc7-923a-b16075641a80 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 913.179817] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 913.179817] env[65385]: value = "resgroup-9" [ 913.179817] env[65385]: _type = "ResourcePool" [ 913.179817] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 913.180230] env[65385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-03eee78e-64f7-46a9-b769-71e30f1e1dac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.211121] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease: (returnval){ [ 913.211121] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0f486-7547-198f-2d7d-17682ed183ec" [ 913.211121] env[65385]: _type = "HttpNfcLease" [ 913.211121] env[65385]: } obtained for vApp import into resource pool (val){ [ 913.211121] env[65385]: value = "resgroup-9" [ 913.211121] env[65385]: _type = "ResourcePool" [ 913.211121] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 913.211487] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the lease: (returnval){ [ 913.211487] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0f486-7547-198f-2d7d-17682ed183ec" [ 913.211487] env[65385]: _type = "HttpNfcLease" [ 913.211487] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 913.217733] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28074998-d5d9-4205-8a06-7b29d73f9597 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.222541] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.222541] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0f486-7547-198f-2d7d-17682ed183ec" [ 913.222541] env[65385]: _type = "HttpNfcLease" [ 913.222541] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 913.228157] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8d7aeb-15cf-406a-b655-0b1f68db7baf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.264426] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f798b892-56b1-4d22-b9bc-025341381158 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.276250] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07bb301-7397-4faf-b938-73a38a14db8f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.293428] env[65385]: DEBUG nova.compute.provider_tree [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.346784] env[65385]: DEBUG nova.network.neutron [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Successfully updated port: 1a20c5c9-1b05-42cf-a71f-894e2921e893 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 913.640998] env[65385]: WARNING neutronclient.v2_0.client [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.641457] env[65385]: WARNING openstack [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.641826] env[65385]: WARNING openstack [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.667896] env[65385]: DEBUG nova.compute.manager [req-5d113fdb-2a18-411c-8b2a-f411bf1e6649 req-68ca9e5e-ab4f-47fe-a342-3562f56b70d2 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Received event network-vif-plugged-1a20c5c9-1b05-42cf-a71f-894e2921e893 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 913.667896] env[65385]: DEBUG oslo_concurrency.lockutils [req-5d113fdb-2a18-411c-8b2a-f411bf1e6649 req-68ca9e5e-ab4f-47fe-a342-3562f56b70d2 service nova] Acquiring lock "fbc67b68-64ed-410a-a43a-efa3f55b3031-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.667896] env[65385]: DEBUG oslo_concurrency.lockutils [req-5d113fdb-2a18-411c-8b2a-f411bf1e6649 req-68ca9e5e-ab4f-47fe-a342-3562f56b70d2 service nova] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.667896] env[65385]: DEBUG oslo_concurrency.lockutils [req-5d113fdb-2a18-411c-8b2a-f411bf1e6649 req-68ca9e5e-ab4f-47fe-a342-3562f56b70d2 service nova] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.667896] env[65385]: DEBUG nova.compute.manager [req-5d113fdb-2a18-411c-8b2a-f411bf1e6649 req-68ca9e5e-ab4f-47fe-a342-3562f56b70d2 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] No waiting events found dispatching network-vif-plugged-1a20c5c9-1b05-42cf-a71f-894e2921e893 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 913.667896] env[65385]: WARNING nova.compute.manager [req-5d113fdb-2a18-411c-8b2a-f411bf1e6649 req-68ca9e5e-ab4f-47fe-a342-3562f56b70d2 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Received unexpected event network-vif-plugged-1a20c5c9-1b05-42cf-a71f-894e2921e893 for instance with vm_state building and task_state spawning. [ 913.720580] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.720580] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0f486-7547-198f-2d7d-17682ed183ec" [ 913.720580] env[65385]: _type = "HttpNfcLease" [ 913.720580] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 913.797560] env[65385]: DEBUG nova.scheduler.client.report [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 913.850549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-fbc67b68-64ed-410a-a43a-efa3f55b3031" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.851077] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-fbc67b68-64ed-410a-a43a-efa3f55b3031" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 913.851077] env[65385]: DEBUG nova.network.neutron [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 914.223014] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.223014] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0f486-7547-198f-2d7d-17682ed183ec" [ 914.223014] env[65385]: _type = "HttpNfcLease" [ 914.223014] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 914.223443] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 914.223443] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a0f486-7547-198f-2d7d-17682ed183ec" [ 914.223443] env[65385]: _type = "HttpNfcLease" [ 914.223443] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 914.224715] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a266f7c9-a6fb-4f19-a0c6-cfac1eb23523 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.239782] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5286929e-dda0-0bc0-ce66-cfa96fdc3142/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 914.239965] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5286929e-dda0-0bc0-ce66-cfa96fdc3142/disk-0.vmdk. {{(pid=65385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 914.304557] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.030s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.309018] env[65385]: WARNING openstack [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.309479] env[65385]: WARNING openstack [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.320582] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.753s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.320582] env[65385]: DEBUG nova.objects.instance [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lazy-loading 'resources' on Instance uuid e02e3848-1b1e-426b-bc0f-24c8e232a9db {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.326414] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1ff16cef-3429-4bf9-84ae-a0e42ad12e94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.357883] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.358606] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.375356] env[65385]: INFO nova.scheduler.client.report [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Deleted allocations for instance 42b9edf6-1873-49c3-8074-8eef654ac371 [ 914.397364] env[65385]: WARNING neutronclient.v2_0.client [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.397364] env[65385]: WARNING openstack [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.397364] env[65385]: WARNING openstack [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.488320] env[65385]: DEBUG nova.network.neutron [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 914.583738] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.584626] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.693539] env[65385]: DEBUG nova.network.neutron [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Updated VIF entry in instance network info cache for port 6785f095-0836-4bc7-923a-b16075641a80. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 914.693928] env[65385]: DEBUG nova.network.neutron [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Updating instance_info_cache with network_info: [{"id": "6785f095-0836-4bc7-923a-b16075641a80", "address": "fa:16:3e:42:7c:73", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6785f095-08", "ovs_interfaceid": "6785f095-0836-4bc7-923a-b16075641a80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.862340] env[65385]: WARNING neutronclient.v2_0.client [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.862340] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.862340] env[65385]: WARNING openstack [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.885471] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1b6efea7-6fc8-4dc6-a8e0-21eb566bc100 tempest-MigrationsAdminTest-1549642071 tempest-MigrationsAdminTest-1549642071-project-member] Lock "42b9edf6-1873-49c3-8074-8eef654ac371" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.416s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.087932] env[65385]: DEBUG nova.network.neutron [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Updating instance_info_cache with network_info: [{"id": "1a20c5c9-1b05-42cf-a71f-894e2921e893", "address": "fa:16:3e:7a:8f:3c", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a20c5c9-1b", "ovs_interfaceid": "1a20c5c9-1b05-42cf-a71f-894e2921e893", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.105779] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Completed reading data from the image iterator. {{(pid=65385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 915.106307] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5286929e-dda0-0bc0-ce66-cfa96fdc3142/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.109052] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7788f4a7-9817-48f4-a5ab-9e12d08dc5e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.127806] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5286929e-dda0-0bc0-ce66-cfa96fdc3142/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.127880] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5286929e-dda0-0bc0-ce66-cfa96fdc3142/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 915.130090] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b6153925-3956-4f99-af95-aa3f4eff04e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.198051] env[65385]: DEBUG oslo_concurrency.lockutils [req-7009317a-6167-4b2c-b7d8-b1af19109ee9 req-c4c227b3-be0b-4876-9c6c-bdc5bbc917f6 service nova] Releasing lock "refresh_cache-ac6fa516-d954-466e-b8e0-e12440492049" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 915.268466] env[65385]: DEBUG nova.compute.manager [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Received event network-changed-2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 915.270863] env[65385]: DEBUG nova.compute.manager [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Refreshing instance network info cache due to event network-changed-2bb8c7e3-459c-4149-8960-0d0df350dc4a. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 915.270863] env[65385]: DEBUG oslo_concurrency.lockutils [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Acquiring lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.270863] env[65385]: DEBUG oslo_concurrency.lockutils [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Acquired lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 915.270863] env[65385]: DEBUG nova.network.neutron [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Refreshing network info cache for port 2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 915.290247] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01579554-3533-4bd6-93f2-2a119221be91 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.299082] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b41d45-0d9d-4c3b-b961-9e3adfb662e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.333446] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad4f24f-f91b-43a9-bcb6-c11140dc8539 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.344063] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6dd8b1-e54f-4e5d-af19-73de992d6303 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.357946] env[65385]: DEBUG nova.compute.provider_tree [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.416055] env[65385]: DEBUG oslo_vmware.rw_handles [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5286929e-dda0-0bc0-ce66-cfa96fdc3142/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 915.416336] env[65385]: INFO nova.virt.vmwareapi.images [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Downloaded image file data 57fab5fe-ac7e-4d91-94ca-0285c2957e33 [ 915.417183] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a43a89e-ee22-41c9-bfed-3febddc8b602 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.434538] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc9ca951-224b-44b4-941f-a785e2f53495 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.462973] env[65385]: INFO nova.virt.vmwareapi.images [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] The imported VM was unregistered [ 915.465892] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Caching image {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 915.466154] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating directory with path [datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33 {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.466441] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20bb6422-b202-40c7-a123-3b599ad9ce6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.480411] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created directory with path [datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33 {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.480676] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3/OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3.vmdk to [datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33/57fab5fe-ac7e-4d91-94ca-0285c2957e33.vmdk. {{(pid=65385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 915.480956] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9dfc32a5-bf4d-4c41-900e-789983d32c8f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.489393] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 915.489393] env[65385]: value = "task-4453957" [ 915.489393] env[65385]: _type = "Task" [ 915.489393] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.502570] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453957, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.596654] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-fbc67b68-64ed-410a-a43a-efa3f55b3031" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 915.598300] env[65385]: DEBUG nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Instance network_info: |[{"id": "1a20c5c9-1b05-42cf-a71f-894e2921e893", "address": "fa:16:3e:7a:8f:3c", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a20c5c9-1b", "ovs_interfaceid": "1a20c5c9-1b05-42cf-a71f-894e2921e893", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 915.598300] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:8f:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a20c5c9-1b05-42cf-a71f-894e2921e893', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.606210] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 915.607166] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.607166] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10696e3d-523d-4866-8f26-8234cf7d8412 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.628135] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.628135] env[65385]: value = "task-4453958" [ 915.628135] env[65385]: _type = "Task" [ 915.628135] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.637240] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453958, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.777360] env[65385]: WARNING neutronclient.v2_0.client [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.778143] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.778591] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.863142] env[65385]: DEBUG nova.scheduler.client.report [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 916.009765] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453957, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.145684] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453958, 'name': CreateVM_Task, 'duration_secs': 0.495045} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.145978] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.146748] env[65385]: WARNING neutronclient.v2_0.client [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 916.147491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.147686] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.148038] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 916.148429] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7365220-a42a-48e6-abdf-dcb01727514d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.157548] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 916.157548] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528725ed-d0fa-ebd8-f8f4-64b46a779318" [ 916.157548] env[65385]: _type = "Task" [ 916.157548] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.175547] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528725ed-d0fa-ebd8-f8f4-64b46a779318, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.286939] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.287648] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.369941] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.050s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.373857] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.086s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.376197] env[65385]: INFO nova.compute.claims [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 916.428606] env[65385]: INFO nova.scheduler.client.report [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Deleted allocations for instance e02e3848-1b1e-426b-bc0f-24c8e232a9db [ 916.507022] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453957, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.540731] env[65385]: WARNING neutronclient.v2_0.client [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 916.540731] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.540731] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.556885] env[65385]: DEBUG nova.compute.manager [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Received event network-changed-1a20c5c9-1b05-42cf-a71f-894e2921e893 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 916.557229] env[65385]: DEBUG nova.compute.manager [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Refreshing instance network info cache due to event network-changed-1a20c5c9-1b05-42cf-a71f-894e2921e893. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 916.557942] env[65385]: DEBUG oslo_concurrency.lockutils [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Acquiring lock "refresh_cache-fbc67b68-64ed-410a-a43a-efa3f55b3031" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.558294] env[65385]: DEBUG oslo_concurrency.lockutils [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Acquired lock "refresh_cache-fbc67b68-64ed-410a-a43a-efa3f55b3031" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.559985] env[65385]: DEBUG nova.network.neutron [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Refreshing network info cache for port 1a20c5c9-1b05-42cf-a71f-894e2921e893 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 916.674081] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528725ed-d0fa-ebd8-f8f4-64b46a779318, 'name': SearchDatastore_Task, 'duration_secs': 0.118666} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.674441] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 916.674686] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.674934] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.675420] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.675420] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.675615] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca999328-64a4-46ec-bc62-96cf9af0c128 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.698756] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.700895] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.700895] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c82d8022-ebbd-486e-a2bd-e3c5ec1b166d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.709895] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 916.709895] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ade7a7-eb46-dd05-d3fc-3a17dc19d25f" [ 916.709895] env[65385]: _type = "Task" [ 916.709895] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.723743] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ade7a7-eb46-dd05-d3fc-3a17dc19d25f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.798141] env[65385]: DEBUG nova.network.neutron [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Updated VIF entry in instance network info cache for port 2bb8c7e3-459c-4149-8960-0d0df350dc4a. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 916.798627] env[65385]: DEBUG nova.network.neutron [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Updating instance_info_cache with network_info: [{"id": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "address": "fa:16:3e:b8:39:d3", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bb8c7e3-45", "ovs_interfaceid": "2bb8c7e3-459c-4149-8960-0d0df350dc4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 916.943487] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e6a1a3ab-7d9a-4242-bcad-5539866f9c58 tempest-ImagesOneServerNegativeTestJSON-1926519048 tempest-ImagesOneServerNegativeTestJSON-1926519048-project-member] Lock "e02e3848-1b1e-426b-bc0f-24c8e232a9db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.904s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.008224] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453957, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.064754] env[65385]: WARNING neutronclient.v2_0.client [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.064754] env[65385]: WARNING openstack [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.064944] env[65385]: WARNING openstack [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.224757] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ade7a7-eb46-dd05-d3fc-3a17dc19d25f, 'name': SearchDatastore_Task, 'duration_secs': 0.088112} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.226115] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33c9f625-6f36-4d84-9afe-8db54f4b4e5a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.236361] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 917.236361] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52123891-c048-280c-08c4-1250ee04846a" [ 917.236361] env[65385]: _type = "Task" [ 917.236361] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.237204] env[65385]: WARNING openstack [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.237574] env[65385]: WARNING openstack [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.269895] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52123891-c048-280c-08c4-1250ee04846a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.302489] env[65385]: DEBUG oslo_concurrency.lockutils [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Releasing lock "refresh_cache-4b0b8859-b63e-4740-bf67-5733c75e1c70" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.302839] env[65385]: DEBUG nova.compute.manager [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 917.303174] env[65385]: DEBUG nova.compute.manager [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing instance network info cache due to event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 917.303313] env[65385]: DEBUG oslo_concurrency.lockutils [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Acquiring lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.303411] env[65385]: DEBUG oslo_concurrency.lockutils [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Acquired lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 917.303586] env[65385]: DEBUG nova.network.neutron [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 917.487747] env[65385]: WARNING neutronclient.v2_0.client [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.489160] env[65385]: WARNING openstack [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.489706] env[65385]: WARNING openstack [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.518488] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453957, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.678936] env[65385]: DEBUG nova.network.neutron [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Updated VIF entry in instance network info cache for port 1a20c5c9-1b05-42cf-a71f-894e2921e893. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 917.683062] env[65385]: DEBUG nova.network.neutron [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Updating instance_info_cache with network_info: [{"id": "1a20c5c9-1b05-42cf-a71f-894e2921e893", "address": "fa:16:3e:7a:8f:3c", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a20c5c9-1b", "ovs_interfaceid": "1a20c5c9-1b05-42cf-a71f-894e2921e893", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 917.767255] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52123891-c048-280c-08c4-1250ee04846a, 'name': SearchDatastore_Task, 'duration_secs': 0.100305} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.768710] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.768999] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] fbc67b68-64ed-410a-a43a-efa3f55b3031/fbc67b68-64ed-410a-a43a-efa3f55b3031.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.770521] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829f4180-533c-4e6d-be68-d5ffc3e72256 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.774503] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5711dd2-5f23-46ea-863c-f9c75375ffbc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.788474] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cda371-b0e3-4e01-a42b-329017b36023 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.792400] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 917.792400] env[65385]: value = "task-4453959" [ 917.792400] env[65385]: _type = "Task" [ 917.792400] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.825243] env[65385]: WARNING neutronclient.v2_0.client [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.826196] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.826338] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.834749] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758858b1-731b-40b9-aba0-4fb3b4454b4f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.842649] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453959, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.846246] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a4533a-cbaf-449f-bc34-319537e7b379 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.861939] env[65385]: DEBUG nova.compute.provider_tree [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.989926] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.990415] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.012564] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453957, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.517254} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.012923] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3/OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3.vmdk to [datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33/57fab5fe-ac7e-4d91-94ca-0285c2957e33.vmdk. [ 918.013093] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Cleaning up location [datastore2] OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 918.013294] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_9225ec3d-fbb0-4e1d-84d4-efd4c6725dc3 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.013580] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f9d5d7e-1605-4e04-aec5-fca40fccef94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.023024] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 918.023024] env[65385]: value = "task-4453960" [ 918.023024] env[65385]: _type = "Task" [ 918.023024] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.031663] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.189138] env[65385]: DEBUG oslo_concurrency.lockutils [req-243f0bc1-5723-4554-94cc-38831e50707b req-de74b096-5ad8-469e-87af-829b599e5712 service nova] Releasing lock "refresh_cache-fbc67b68-64ed-410a-a43a-efa3f55b3031" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 918.310913] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453959, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.371295] env[65385]: DEBUG nova.scheduler.client.report [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 918.513243] env[65385]: WARNING neutronclient.v2_0.client [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.513968] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.514367] env[65385]: WARNING openstack [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.531899] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040775} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.532258] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.532361] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33/57fab5fe-ac7e-4d91-94ca-0285c2957e33.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 918.532607] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33/57fab5fe-ac7e-4d91-94ca-0285c2957e33.vmdk to [datastore2] ac6fa516-d954-466e-b8e0-e12440492049/ac6fa516-d954-466e-b8e0-e12440492049.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 918.532868] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7509ac5-2226-49e6-a1a8-b26cbe64cce0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.541099] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 918.541099] env[65385]: value = "task-4453961" [ 918.541099] env[65385]: _type = "Task" [ 918.541099] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.549409] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453961, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.809365] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453959, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.633196} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.809752] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] fbc67b68-64ed-410a-a43a-efa3f55b3031/fbc67b68-64ed-410a-a43a-efa3f55b3031.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.810041] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.810390] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ccc8aaf-4f40-4b39-ade1-7ca643b42bda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.819745] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 918.819745] env[65385]: value = "task-4453962" [ 918.819745] env[65385]: _type = "Task" [ 918.819745] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.831494] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453962, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.877017] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 918.877641] env[65385]: DEBUG nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 918.880785] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.702s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.884434] env[65385]: DEBUG nova.objects.instance [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lazy-loading 'resources' on Instance uuid eb9d0510-b453-4695-9e1d-731217b9f8ec {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.063882] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453961, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.109301] env[65385]: DEBUG nova.network.neutron [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updated VIF entry in instance network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 919.109939] env[65385]: DEBUG nova.network.neutron [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 919.337108] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453962, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082917} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.337525] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.338404] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cc6e4b-8cab-4077-ba3f-1c7e0f1ac810 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.370762] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] fbc67b68-64ed-410a-a43a-efa3f55b3031/fbc67b68-64ed-410a-a43a-efa3f55b3031.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.371134] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cd307aa-15ae-4c17-8a2e-b74b7b04f6e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.389678] env[65385]: DEBUG nova.compute.utils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 919.393610] env[65385]: DEBUG nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 919.393825] env[65385]: DEBUG nova.network.neutron [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 919.394222] env[65385]: WARNING neutronclient.v2_0.client [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.394532] env[65385]: WARNING neutronclient.v2_0.client [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.395209] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.395504] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 919.416195] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 919.416195] env[65385]: value = "task-4453963" [ 919.416195] env[65385]: _type = "Task" [ 919.416195] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.430306] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453963, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.518280] env[65385]: DEBUG nova.policy [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1326ee8a1084b738e7bcce0c04d0d40', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e073e1e0d8a045a48ea8fe3c4dd28b6f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 919.554249] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453961, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.614247] env[65385]: DEBUG oslo_concurrency.lockutils [req-268144bf-568e-4caa-9137-50e1e942f014 req-864daaab-a885-47d2-a38d-ed4bc071471d service nova] Releasing lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.769767] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57a8e25-c376-49eb-9255-ab4a1e6bd8f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.780013] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51296dde-71f3-49cd-a440-6fc6b36f2f90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.815956] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8de6ee-dee8-4f01-8fc2-5b8236168541 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.826267] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d141e6-8040-44ca-b3fa-61b15180b86c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.846514] env[65385]: DEBUG nova.compute.provider_tree [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.894666] env[65385]: DEBUG nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 919.929716] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453963, 'name': ReconfigVM_Task, 'duration_secs': 0.493659} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.929936] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Reconfigured VM instance instance-0000004b to attach disk [datastore2] fbc67b68-64ed-410a-a43a-efa3f55b3031/fbc67b68-64ed-410a-a43a-efa3f55b3031.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.932016] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d20e4956-9594-46dc-85c5-c349ebba0016 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.940907] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 919.940907] env[65385]: value = "task-4453964" [ 919.940907] env[65385]: _type = "Task" [ 919.940907] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.951245] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453964, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.019428] env[65385]: DEBUG nova.network.neutron [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Successfully created port: 984014f3-cd87-4054-a971-942d597305f7 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 920.059116] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453961, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.349790] env[65385]: DEBUG nova.scheduler.client.report [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 920.456494] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453964, 'name': Rename_Task, 'duration_secs': 0.188744} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.456824] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.457258] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05ff7186-e7f6-4406-87b5-ed51772d0da2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.467628] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 920.467628] env[65385]: value = "task-4453965" [ 920.467628] env[65385]: _type = "Task" [ 920.467628] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.479803] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453965, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.557023] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453961, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.859163] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.861050] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.756s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.862961] env[65385]: INFO nova.compute.claims [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.894896] env[65385]: INFO nova.scheduler.client.report [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Deleted allocations for instance eb9d0510-b453-4695-9e1d-731217b9f8ec [ 920.909879] env[65385]: DEBUG nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 920.958232] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 920.958557] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 920.958747] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 920.958956] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 920.959150] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 920.960152] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 920.960152] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.960152] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 920.960152] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 920.960352] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 920.960450] env[65385]: DEBUG nova.virt.hardware [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 920.962121] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a29fe4-d4da-4e8e-bbfb-f56664a43c06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.977468] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25306c8-176c-406b-abfd-f755a4876ca0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.999686] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453965, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.056290] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453961, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.432426} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.056697] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/57fab5fe-ac7e-4d91-94ca-0285c2957e33/57fab5fe-ac7e-4d91-94ca-0285c2957e33.vmdk to [datastore2] ac6fa516-d954-466e-b8e0-e12440492049/ac6fa516-d954-466e-b8e0-e12440492049.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.057420] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a99287-be02-4f19-b07c-2b60ad2ca937 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.091933] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] ac6fa516-d954-466e-b8e0-e12440492049/ac6fa516-d954-466e-b8e0-e12440492049.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.094440] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c0dc260-9bd2-4f7b-8cb2-b850e239b906 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.118982] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 921.118982] env[65385]: value = "task-4453966" [ 921.118982] env[65385]: _type = "Task" [ 921.118982] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.129034] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453966, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.404310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5219133a-e9c6-4078-8fc1-bc203ca91b76 tempest-ServersAdminTestJSON-1855485524 tempest-ServersAdminTestJSON-1855485524-project-member] Lock "eb9d0510-b453-4695-9e1d-731217b9f8ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.764s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.486072] env[65385]: DEBUG oslo_vmware.api [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453965, 'name': PowerOnVM_Task, 'duration_secs': 0.544272} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.486353] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.486547] env[65385]: INFO nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Took 9.18 seconds to spawn the instance on the hypervisor. [ 921.486725] env[65385]: DEBUG nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 921.487636] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac07fc81-db82-4f78-9eb4-58f7852a0395 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.632794] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453966, 'name': ReconfigVM_Task, 'duration_secs': 0.295745} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.633116] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Reconfigured VM instance instance-0000004a to attach disk [datastore2] ac6fa516-d954-466e-b8e0-e12440492049/ac6fa516-d954-466e-b8e0-e12440492049.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.633834] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8de7591a-5ea3-4a76-95d6-b51ee9b8af34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.642838] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 921.642838] env[65385]: value = "task-4453967" [ 921.642838] env[65385]: _type = "Task" [ 921.642838] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.652783] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453967, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.700688] env[65385]: DEBUG nova.network.neutron [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Successfully updated port: 984014f3-cd87-4054-a971-942d597305f7 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 922.018888] env[65385]: INFO nova.compute.manager [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Took 36.94 seconds to build instance. [ 922.101497] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.102155] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.164190] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453967, 'name': Rename_Task, 'duration_secs': 0.148259} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.167088] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.167088] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-665bf363-b037-4e55-ad4e-081d53bbe08e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.174465] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 922.174465] env[65385]: value = "task-4453968" [ 922.174465] env[65385]: _type = "Task" [ 922.174465] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.188218] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453968, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.190090] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70cf404a-891f-4546-a9a2-fe100d752098 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.198196] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5534e81-981b-45ab-b090-92b8e02aca3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.202889] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "refresh_cache-e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.203735] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquired lock "refresh_cache-e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.203978] env[65385]: DEBUG nova.network.neutron [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 922.243288] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37293c17-cb21-4369-9677-43af25c1112e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.251876] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521bf0db-1520-49c9-91a9-8d202c9f12da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.267209] env[65385]: DEBUG nova.compute.provider_tree [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.523175] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0720bbe7-92f5-46e3-94ec-445b64e969ec tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.462s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 922.605691] env[65385]: DEBUG nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 922.664854] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-c18b5bde-1a17-4954-a2c5-582815ebfec8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.665154] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-c18b5bde-1a17-4954-a2c5-582815ebfec8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.665567] env[65385]: DEBUG nova.objects.instance [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'flavor' on Instance uuid c18b5bde-1a17-4954-a2c5-582815ebfec8 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.687277] env[65385]: DEBUG oslo_vmware.api [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453968, 'name': PowerOnVM_Task, 'duration_secs': 0.48449} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.687637] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 922.687737] env[65385]: INFO nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Took 15.42 seconds to spawn the instance on the hypervisor. [ 922.687913] env[65385]: DEBUG nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 922.688958] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e21c31-ca51-4ae5-ab90-3f7b49a9f11b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.707745] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.708280] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.770027] env[65385]: DEBUG nova.scheduler.client.report [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 922.917362] env[65385]: DEBUG nova.network.neutron [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 922.926266] env[65385]: DEBUG nova.compute.manager [req-447b427d-7917-442a-ad1d-3015f5884568 req-20f09953-a5b7-4a3c-bf59-34b875344aaa service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Received event network-vif-plugged-984014f3-cd87-4054-a971-942d597305f7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 922.926519] env[65385]: DEBUG oslo_concurrency.lockutils [req-447b427d-7917-442a-ad1d-3015f5884568 req-20f09953-a5b7-4a3c-bf59-34b875344aaa service nova] Acquiring lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.926748] env[65385]: DEBUG oslo_concurrency.lockutils [req-447b427d-7917-442a-ad1d-3015f5884568 req-20f09953-a5b7-4a3c-bf59-34b875344aaa service nova] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.926918] env[65385]: DEBUG oslo_concurrency.lockutils [req-447b427d-7917-442a-ad1d-3015f5884568 req-20f09953-a5b7-4a3c-bf59-34b875344aaa service nova] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 922.927089] env[65385]: DEBUG nova.compute.manager [req-447b427d-7917-442a-ad1d-3015f5884568 req-20f09953-a5b7-4a3c-bf59-34b875344aaa service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] No waiting events found dispatching network-vif-plugged-984014f3-cd87-4054-a971-942d597305f7 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 922.927353] env[65385]: WARNING nova.compute.manager [req-447b427d-7917-442a-ad1d-3015f5884568 req-20f09953-a5b7-4a3c-bf59-34b875344aaa service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Received unexpected event network-vif-plugged-984014f3-cd87-4054-a971-942d597305f7 for instance with vm_state building and task_state spawning. [ 922.950837] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.951504] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.150450] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 923.170318] env[65385]: WARNING neutronclient.v2_0.client [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.171098] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.171443] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.207681] env[65385]: INFO nova.compute.manager [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Took 41.14 seconds to build instance. [ 923.275670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.276262] env[65385]: DEBUG nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 923.279077] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.146s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 923.282371] env[65385]: DEBUG nova.objects.instance [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lazy-loading 'resources' on Instance uuid fa9a5278-1477-485a-9201-a37187488aef {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.360116] env[65385]: WARNING neutronclient.v2_0.client [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.360116] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.360116] env[65385]: WARNING openstack [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.512141] env[65385]: DEBUG nova.objects.instance [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'pci_requests' on Instance uuid c18b5bde-1a17-4954-a2c5-582815ebfec8 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.663535] env[65385]: DEBUG nova.network.neutron [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Updating instance_info_cache with network_info: [{"id": "984014f3-cd87-4054-a971-942d597305f7", "address": "fa:16:3e:a9:75:3f", "network": {"id": "bf876837-9b57-48eb-a272-00860a734464", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1389481625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e073e1e0d8a045a48ea8fe3c4dd28b6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984014f3-cd", "ovs_interfaceid": "984014f3-cd87-4054-a971-942d597305f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 923.710669] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1552767e-ab6e-4bc4-b971-b88fc706723f tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "ac6fa516-d954-466e-b8e0-e12440492049" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.651s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.788332] env[65385]: DEBUG nova.compute.utils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 923.791133] env[65385]: DEBUG nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 923.791229] env[65385]: DEBUG nova.network.neutron [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 923.791525] env[65385]: WARNING neutronclient.v2_0.client [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.791839] env[65385]: WARNING neutronclient.v2_0.client [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.792469] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.792825] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.802905] env[65385]: DEBUG nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 924.012938] env[65385]: DEBUG nova.policy [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c44b5bf353f4d93bd572b301253d1b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '02f51c46458e425d9d0c330c02bde184', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 924.018434] env[65385]: DEBUG nova.objects.base [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 924.018630] env[65385]: DEBUG nova.network.neutron [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 924.018941] env[65385]: WARNING neutronclient.v2_0.client [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.019239] env[65385]: WARNING neutronclient.v2_0.client [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.020260] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.020260] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.054696] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd6363f-011b-4a45-96bf-b0d73becc1e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.063889] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c0edbb-5dfb-4d0f-9a35-417a6bef7e8a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.099457] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b75351f-8f0d-4870-ad69-adcf3f1cec87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.108285] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8c672f-bc3e-4698-b92e-6f1db35e1291 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.122931] env[65385]: DEBUG nova.compute.provider_tree [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.130018] env[65385]: DEBUG nova.policy [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 924.166594] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Releasing lock "refresh_cache-e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 924.166878] env[65385]: DEBUG nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Instance network_info: |[{"id": "984014f3-cd87-4054-a971-942d597305f7", "address": "fa:16:3e:a9:75:3f", "network": {"id": "bf876837-9b57-48eb-a272-00860a734464", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1389481625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e073e1e0d8a045a48ea8fe3c4dd28b6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984014f3-cd", "ovs_interfaceid": "984014f3-cd87-4054-a971-942d597305f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 924.167756] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:75:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '984014f3-cd87-4054-a971-942d597305f7', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.176106] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Creating folder: Project (e073e1e0d8a045a48ea8fe3c4dd28b6f). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.176439] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d2cf909-1918-4927-9e5c-290af5c05d52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.188577] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Created folder: Project (e073e1e0d8a045a48ea8fe3c4dd28b6f) in parent group-v870881. [ 924.188765] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Creating folder: Instances. Parent ref: group-v871096. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 924.189055] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7ffb3c5-5bd5-4d50-82a6-4ed25411eafa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.200299] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Created folder: Instances in parent group-v871096. [ 924.201029] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 924.201029] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 924.201029] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b24187b-6c86-49e0-b512-7811e1725e90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.225511] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.225511] env[65385]: value = "task-4453971" [ 924.225511] env[65385]: _type = "Task" [ 924.225511] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.240103] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453971, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.544724] env[65385]: DEBUG nova.network.neutron [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Successfully created port: 37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 924.581040] env[65385]: DEBUG nova.network.neutron [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Successfully created port: 2ff7c104-f798-4d67-b92b-3d73b0810a38 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 924.633864] env[65385]: DEBUG nova.scheduler.client.report [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 924.736756] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453971, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.813173] env[65385]: DEBUG nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 924.840401] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 924.840659] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 924.840840] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 924.841048] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 924.841200] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 924.841346] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 924.841559] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.841713] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 924.841895] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 924.842113] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 924.842292] env[65385]: DEBUG nova.virt.hardware [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 924.843298] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9792614-53b2-4c02-af4e-368924b4b167 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.853025] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fb150f-df37-48d6-bb76-ad9ede92ea5b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.141768] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.144604] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.516s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.144909] env[65385]: DEBUG nova.objects.instance [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lazy-loading 'resources' on Instance uuid 6dae15ad-a340-47e3-ab6b-95b25043ca27 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.163161] env[65385]: INFO nova.scheduler.client.report [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Deleted allocations for instance fa9a5278-1477-485a-9201-a37187488aef [ 925.241270] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453971, 'name': CreateVM_Task, 'duration_secs': 0.525469} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.241270] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.242028] env[65385]: WARNING neutronclient.v2_0.client [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.242319] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.242473] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.242806] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 925.243402] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdf020df-e2ca-4f31-a82e-7ebbe7fe5ae4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.249453] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 925.249453] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52160aef-3c42-e178-8a45-82959cf77b94" [ 925.249453] env[65385]: _type = "Task" [ 925.249453] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.261107] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52160aef-3c42-e178-8a45-82959cf77b94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.673343] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e37f739-307d-4f37-982d-5e60d85a3881 tempest-ServerMetadataNegativeTestJSON-686171521 tempest-ServerMetadataNegativeTestJSON-686171521-project-member] Lock "fa9a5278-1477-485a-9201-a37187488aef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.122s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.759473] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52160aef-3c42-e178-8a45-82959cf77b94, 'name': SearchDatastore_Task, 'duration_secs': 0.01513} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.762319] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.762548] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.762773] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.762917] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.763099] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.763887] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94686c84-7d30-4cab-8044-1ffad9049201 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.776011] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.776197] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.776905] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2428e277-e1d8-4280-8430-832a1b3de82b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.782239] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 925.782239] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52786873-e291-8a27-9b05-48d6ab96d4f6" [ 925.782239] env[65385]: _type = "Task" [ 925.782239] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.791582] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52786873-e291-8a27-9b05-48d6ab96d4f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.885362] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83603bcc-1aae-4ae8-8a9d-27f1d9134194 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.893459] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00721ddc-fe1e-4233-ba78-4a68e48b0b49 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.925923] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b28931-5b88-4b90-8302-d4637cda15c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.934715] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed78a5b1-cc55-4efa-a2bb-92475dcc80c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.949987] env[65385]: DEBUG nova.compute.provider_tree [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.176461] env[65385]: DEBUG nova.network.neutron [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Successfully updated port: 37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 926.296557] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52786873-e291-8a27-9b05-48d6ab96d4f6, 'name': SearchDatastore_Task, 'duration_secs': 0.028782} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.297532] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9984c882-3020-4475-8236-56692077a32f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.304381] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 926.304381] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520e1417-3343-fda6-7bac-57815ddd40df" [ 926.304381] env[65385]: _type = "Task" [ 926.304381] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.314344] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520e1417-3343-fda6-7bac-57815ddd40df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.455696] env[65385]: DEBUG nova.scheduler.client.report [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 926.547409] env[65385]: DEBUG nova.network.neutron [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Successfully updated port: 2ff7c104-f798-4d67-b92b-3d73b0810a38 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 926.681105] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.681105] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquired lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 926.681105] env[65385]: DEBUG nova.network.neutron [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 926.821401] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520e1417-3343-fda6-7bac-57815ddd40df, 'name': SearchDatastore_Task, 'duration_secs': 0.022788} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.821860] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.822196] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e8f9418e-904c-4e3f-8ea9-91ab0fdd080d/e8f9418e-904c-4e3f-8ea9-91ab0fdd080d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.822535] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dcc0f58c-c641-4f1a-b82f-2dcfb8b7eb5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.831583] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 926.831583] env[65385]: value = "task-4453972" [ 926.831583] env[65385]: _type = "Task" [ 926.831583] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.841976] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.959962] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.962730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.854s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.964455] env[65385]: INFO nova.compute.claims [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 927.010229] env[65385]: INFO nova.scheduler.client.report [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Deleted allocations for instance 6dae15ad-a340-47e3-ab6b-95b25043ca27 [ 927.050377] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.050569] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.050929] env[65385]: DEBUG nova.network.neutron [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 927.183982] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.184426] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.327637] env[65385]: DEBUG nova.network.neutron [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 927.344764] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453972, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.361634] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.361763] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.422032] env[65385]: WARNING neutronclient.v2_0.client [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.423176] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.423176] env[65385]: WARNING openstack [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.521085] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0b62b813-081b-497c-bfd0-6edf7c788da5 tempest-ServerTagsTestJSON-645774926 tempest-ServerTagsTestJSON-645774926-project-member] Lock "6dae15ad-a340-47e3-ab6b-95b25043ca27" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.350s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.554618] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.554958] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.656152] env[65385]: DEBUG nova.network.neutron [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Updating instance_info_cache with network_info: [{"id": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "address": "fa:16:3e:e2:4b:e7", "network": {"id": "8f9a0fa3-6de0-4601-bed3-6ba0670da6e1", "bridge": "br-int", "label": "tempest-ServersTestJSON-1442256052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "02f51c46458e425d9d0c330c02bde184", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0d7a2b2f-3b49-4dc8-9096-af16144b27a9", "external-id": "nsx-vlan-transportzone-492", "segmentation_id": 492, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37cdfc1d-4b", "ovs_interfaceid": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 927.691352] env[65385]: WARNING nova.network.neutron [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] 16967e55-c149-4a58-905f-c47b7facac11 already exists in list: networks containing: ['16967e55-c149-4a58-905f-c47b7facac11']. ignoring it [ 927.741748] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.742492] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.807061] env[65385]: WARNING neutronclient.v2_0.client [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.807740] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.808095] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.845105] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523692} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.845347] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e8f9418e-904c-4e3f-8ea9-91ab0fdd080d/e8f9418e-904c-4e3f-8ea9-91ab0fdd080d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 927.845696] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.845795] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5bb3c998-5c30-4f94-a1f1-26a9c91877bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.853152] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 927.853152] env[65385]: value = "task-4453973" [ 927.853152] env[65385]: _type = "Task" [ 927.853152] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.867764] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.900888] env[65385]: DEBUG nova.compute.manager [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Received event network-changed-984014f3-cd87-4054-a971-942d597305f7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 927.900888] env[65385]: DEBUG nova.compute.manager [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Refreshing instance network info cache due to event network-changed-984014f3-cd87-4054-a971-942d597305f7. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 927.900981] env[65385]: DEBUG oslo_concurrency.lockutils [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Acquiring lock "refresh_cache-e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.901399] env[65385]: DEBUG oslo_concurrency.lockutils [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Acquired lock "refresh_cache-e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.901399] env[65385]: DEBUG nova.network.neutron [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Refreshing network info cache for port 984014f3-cd87-4054-a971-942d597305f7 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 927.918511] env[65385]: DEBUG nova.compute.manager [req-70516100-fdbb-4b95-8d3d-56a8d024aa13 req-8917d121-e389-4049-915d-2d1e896fe2e0 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Received event network-vif-plugged-37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 927.918746] env[65385]: DEBUG oslo_concurrency.lockutils [req-70516100-fdbb-4b95-8d3d-56a8d024aa13 req-8917d121-e389-4049-915d-2d1e896fe2e0 service nova] Acquiring lock "3febf6af-699a-4fa1-b079-3790cd1095e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.918973] env[65385]: DEBUG oslo_concurrency.lockutils [req-70516100-fdbb-4b95-8d3d-56a8d024aa13 req-8917d121-e389-4049-915d-2d1e896fe2e0 service nova] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.919165] env[65385]: DEBUG oslo_concurrency.lockutils [req-70516100-fdbb-4b95-8d3d-56a8d024aa13 req-8917d121-e389-4049-915d-2d1e896fe2e0 service nova] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.920072] env[65385]: DEBUG nova.compute.manager [req-70516100-fdbb-4b95-8d3d-56a8d024aa13 req-8917d121-e389-4049-915d-2d1e896fe2e0 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] No waiting events found dispatching network-vif-plugged-37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 927.920072] env[65385]: WARNING nova.compute.manager [req-70516100-fdbb-4b95-8d3d-56a8d024aa13 req-8917d121-e389-4049-915d-2d1e896fe2e0 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Received unexpected event network-vif-plugged-37cdfc1d-4b24-4ae6-ac71-e0d63602b872 for instance with vm_state building and task_state spawning. [ 927.935673] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "ac6fa516-d954-466e-b8e0-e12440492049" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.935928] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "ac6fa516-d954-466e-b8e0-e12440492049" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.936144] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "ac6fa516-d954-466e-b8e0-e12440492049-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.936314] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "ac6fa516-d954-466e-b8e0-e12440492049-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.936471] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "ac6fa516-d954-466e-b8e0-e12440492049-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.938499] env[65385]: INFO nova.compute.manager [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Terminating instance [ 927.943725] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.944106] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.035511] env[65385]: WARNING neutronclient.v2_0.client [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.036332] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.037184] env[65385]: WARNING openstack [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.155889] env[65385]: DEBUG nova.network.neutron [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updating instance_info_cache with network_info: [{"id": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "address": "fa:16:3e:42:06:97", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d96802a-11", "ovs_interfaceid": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2ff7c104-f798-4d67-b92b-3d73b0810a38", "address": "fa:16:3e:e8:96:84", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff7c104-f7", "ovs_interfaceid": "2ff7c104-f798-4d67-b92b-3d73b0810a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 928.161041] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Releasing lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.162220] env[65385]: DEBUG nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Instance network_info: |[{"id": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "address": "fa:16:3e:e2:4b:e7", "network": {"id": "8f9a0fa3-6de0-4601-bed3-6ba0670da6e1", "bridge": "br-int", "label": "tempest-ServersTestJSON-1442256052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "02f51c46458e425d9d0c330c02bde184", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0d7a2b2f-3b49-4dc8-9096-af16144b27a9", "external-id": "nsx-vlan-transportzone-492", "segmentation_id": 492, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37cdfc1d-4b", "ovs_interfaceid": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 928.164484] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:4b:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0d7a2b2f-3b49-4dc8-9096-af16144b27a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37cdfc1d-4b24-4ae6-ac71-e0d63602b872', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.173960] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Creating folder: Project (02f51c46458e425d9d0c330c02bde184). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 928.178560] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee164e9f-811f-440b-a8b5-6a8ea55bdfa9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.192084] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Created folder: Project (02f51c46458e425d9d0c330c02bde184) in parent group-v870881. [ 928.192306] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Creating folder: Instances. Parent ref: group-v871099. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 928.192564] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b399a0a3-48f7-4523-a872-3022b6a65e95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.207687] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Created folder: Instances in parent group-v871099. [ 928.208142] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 928.213365] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.213651] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-693564ef-a31d-49cf-b580-77d4ce859cf8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.238014] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.238014] env[65385]: value = "task-4453976" [ 928.238014] env[65385]: _type = "Task" [ 928.238014] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.246558] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453976, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.271659] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb796ced-2ca0-4d58-81c1-6d9c21a80f2e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.282792] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9bbb55-c64b-4920-87a1-9dbcab2fc3df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.324653] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7380194a-396a-42b8-8e78-2b2d1770a5ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.335321] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f392ddd2-8ada-4d87-89f5-dfee632d700e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.351424] env[65385]: DEBUG nova.compute.provider_tree [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.368815] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075261} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.369098] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.369918] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e833a60b-a4a6-40ce-a979-4a0fbc4a4e50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.394418] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] e8f9418e-904c-4e3f-8ea9-91ab0fdd080d/e8f9418e-904c-4e3f-8ea9-91ab0fdd080d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.394755] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f923328-abe8-4f83-826c-6281e6ce9e92 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.411194] env[65385]: WARNING neutronclient.v2_0.client [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.411896] env[65385]: WARNING openstack [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.412299] env[65385]: WARNING openstack [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.425152] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 928.425152] env[65385]: value = "task-4453977" [ 928.425152] env[65385]: _type = "Task" [ 928.425152] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.434415] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453977, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.444243] env[65385]: DEBUG nova.compute.manager [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 928.444243] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 928.445550] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43764b9-9df9-4024-b138-8c5011d697b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.454097] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.454254] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98c71c1c-7d65-4387-8835-48f2ab85e52b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.461529] env[65385]: DEBUG oslo_vmware.api [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 928.461529] env[65385]: value = "task-4453978" [ 928.461529] env[65385]: _type = "Task" [ 928.461529] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.471030] env[65385]: DEBUG oslo_vmware.api [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.505196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.505477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.506728] env[65385]: DEBUG nova.compute.manager [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 928.506880] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c8511c-867e-402e-adde-d8773ba67f3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.514872] env[65385]: DEBUG nova.compute.manager [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 928.515480] env[65385]: DEBUG nova.objects.instance [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'flavor' on Instance uuid fbc67b68-64ed-410a-a43a-efa3f55b3031 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.665487] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.666438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.666590] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 928.673161] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980419e8-c5fd-4ebb-bd98-18de9870c561 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.693564] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 928.693857] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 928.694084] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 928.694346] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 928.694483] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 928.695315] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 928.695315] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.695315] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 928.695559] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 928.695658] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 928.696322] env[65385]: DEBUG nova.virt.hardware [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 928.703129] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Reconfiguring VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 928.703350] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa9521af-29ba-48f3-9c5d-62eb5df93bc9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.723262] env[65385]: DEBUG oslo_vmware.api [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 928.723262] env[65385]: value = "task-4453979" [ 928.723262] env[65385]: _type = "Task" [ 928.723262] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.740017] env[65385]: DEBUG oslo_vmware.api [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453979, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.751289] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453976, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.806822] env[65385]: WARNING openstack [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.807301] env[65385]: WARNING openstack [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.854748] env[65385]: DEBUG nova.scheduler.client.report [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 928.885950] env[65385]: WARNING neutronclient.v2_0.client [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.886639] env[65385]: WARNING openstack [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.886978] env[65385]: WARNING openstack [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.936846] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453977, 'name': ReconfigVM_Task, 'duration_secs': 0.416571} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.942789] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Reconfigured VM instance instance-0000004c to attach disk [datastore2] e8f9418e-904c-4e3f-8ea9-91ab0fdd080d/e8f9418e-904c-4e3f-8ea9-91ab0fdd080d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.943804] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79ac1c5a-d9df-4308-b583-5a7371efb1d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.953218] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 928.953218] env[65385]: value = "task-4453980" [ 928.953218] env[65385]: _type = "Task" [ 928.953218] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.963910] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453980, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.973211] env[65385]: DEBUG oslo_vmware.api [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453978, 'name': PowerOffVM_Task, 'duration_secs': 0.306077} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.973531] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.973740] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 928.974020] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-052630e6-8a1e-478f-8b65-387c047c741e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.985379] env[65385]: DEBUG nova.network.neutron [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Updated VIF entry in instance network info cache for port 984014f3-cd87-4054-a971-942d597305f7. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 928.985739] env[65385]: DEBUG nova.network.neutron [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Updating instance_info_cache with network_info: [{"id": "984014f3-cd87-4054-a971-942d597305f7", "address": "fa:16:3e:a9:75:3f", "network": {"id": "bf876837-9b57-48eb-a272-00860a734464", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1389481625-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e073e1e0d8a045a48ea8fe3c4dd28b6f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap984014f3-cd", "ovs_interfaceid": "984014f3-cd87-4054-a971-942d597305f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 929.048176] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.048413] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.048608] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleting the datastore file [datastore2] ac6fa516-d954-466e-b8e0-e12440492049 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.048901] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e6993c6-8ac7-4521-afc6-1b53a0f67dae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.056747] env[65385]: DEBUG oslo_vmware.api [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 929.056747] env[65385]: value = "task-4453982" [ 929.056747] env[65385]: _type = "Task" [ 929.056747] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.068303] env[65385]: DEBUG oslo_vmware.api [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453982, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.235891] env[65385]: DEBUG oslo_vmware.api [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453979, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.249952] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453976, 'name': CreateVM_Task, 'duration_secs': 0.529325} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.250355] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 929.250929] env[65385]: WARNING neutronclient.v2_0.client [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.251463] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.251644] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.252038] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 929.252391] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68b0c080-1737-4e5d-8a6d-6ea247b6b8d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.258849] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 929.258849] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d39c50-07cd-52e5-7e38-2f88b12a24af" [ 929.258849] env[65385]: _type = "Task" [ 929.258849] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.269790] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d39c50-07cd-52e5-7e38-2f88b12a24af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.360494] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.361146] env[65385]: DEBUG nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 929.364024] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.182s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.364242] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.367461] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.389s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.367730] env[65385]: DEBUG nova.objects.instance [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lazy-loading 'resources' on Instance uuid cf37e10a-14f4-4456-bad0-d7528457658d {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.416489] env[65385]: INFO nova.scheduler.client.report [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted allocations for instance 89f3c815-8671-47ce-9e74-bf6e652bb3c3 [ 929.467344] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453980, 'name': Rename_Task, 'duration_secs': 0.175058} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.467904] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.468711] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cc71073-1af2-41d9-a6a8-971f7858765f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.480715] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 929.480715] env[65385]: value = "task-4453983" [ 929.480715] env[65385]: _type = "Task" [ 929.480715] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.489861] env[65385]: DEBUG oslo_concurrency.lockutils [req-0a739f54-e776-4e84-b729-085e8094d12a req-21da7c36-4ce5-4c78-add8-b197a63335b6 service nova] Releasing lock "refresh_cache-e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.497321] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.527804] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.527804] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-641496df-bf44-4f13-9194-c21ab6665f4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.534923] env[65385]: DEBUG oslo_vmware.api [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 929.534923] env[65385]: value = "task-4453984" [ 929.534923] env[65385]: _type = "Task" [ 929.534923] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.547043] env[65385]: DEBUG oslo_vmware.api [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453984, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.571564] env[65385]: DEBUG oslo_vmware.api [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4453982, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158199} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.571872] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.572218] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.572461] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.572709] env[65385]: INFO nova.compute.manager [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Took 1.13 seconds to destroy the instance on the hypervisor. [ 929.572979] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 929.574081] env[65385]: DEBUG nova.compute.manager [-] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 929.574081] env[65385]: DEBUG nova.network.neutron [-] [instance: ac6fa516-d954-466e-b8e0-e12440492049] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 929.574081] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.574467] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.574822] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.734863] env[65385]: DEBUG oslo_vmware.api [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453979, 'name': ReconfigVM_Task, 'duration_secs': 0.703757} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.735373] env[65385]: WARNING neutronclient.v2_0.client [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.735702] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.735987] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Reconfigured VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 929.771525] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d39c50-07cd-52e5-7e38-2f88b12a24af, 'name': SearchDatastore_Task, 'duration_secs': 0.012345} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.771829] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.772086] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.772295] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.772433] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.772605] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.773237] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32786e26-1ddd-4cdb-9c81-6cec037db82d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.784351] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.785293] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.785293] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cda572eb-98c7-4559-9430-15bc858b46b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.791812] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 929.791812] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524687c5-9028-491d-a6d3-ec33a0154196" [ 929.791812] env[65385]: _type = "Task" [ 929.791812] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.801358] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524687c5-9028-491d-a6d3-ec33a0154196, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.870670] env[65385]: DEBUG nova.compute.utils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 929.874920] env[65385]: DEBUG nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 929.911445] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.927294] env[65385]: DEBUG oslo_concurrency.lockutils [None req-62f134da-5c66-4e8a-b4d5-d7c2e54f8f38 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "89f3c815-8671-47ce-9e74-bf6e652bb3c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.417s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.993753] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453983, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.044765] env[65385]: DEBUG oslo_vmware.api [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453984, 'name': PowerOffVM_Task, 'duration_secs': 0.292345} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.045388] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.045646] env[65385]: DEBUG nova.compute.manager [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 930.051142] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcab1a70-ad8f-4b27-b9dd-fd749e3dcc65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.145505] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3e71bf-3e60-49bb-96f4-2c46312a46a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.155461] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2407e7-31b5-411d-b0ee-9212de630a50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.197420] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1449f74-e11c-4cff-938e-4e22896274a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.207450] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f6e276-7d62-48b0-b3cf-0c19f0a9fa62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.222947] env[65385]: DEBUG nova.compute.provider_tree [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.243994] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4f8158e-162e-4967-86f3-87f1eaee2961 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-c18b5bde-1a17-4954-a2c5-582815ebfec8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.579s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.305821] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524687c5-9028-491d-a6d3-ec33a0154196, 'name': SearchDatastore_Task, 'duration_secs': 0.021942} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.308022] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4987d3bb-129e-4d7d-a251-952384efa55e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.316997] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 930.316997] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525364ff-ffc6-55f8-0a3f-45e1ca79ce18" [ 930.316997] env[65385]: _type = "Task" [ 930.316997] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.325015] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525364ff-ffc6-55f8-0a3f-45e1ca79ce18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.377405] env[65385]: DEBUG nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 930.497072] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453983, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.564353] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2a50b06d-1aba-4414-97eb-3d256f8434f6 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.059s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.728644] env[65385]: DEBUG nova.scheduler.client.report [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 930.753268] env[65385]: DEBUG nova.network.neutron [-] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 930.828283] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525364ff-ffc6-55f8-0a3f-45e1ca79ce18, 'name': SearchDatastore_Task, 'duration_secs': 0.015099} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.828572] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 930.828839] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 3febf6af-699a-4fa1-b079-3790cd1095e5/3febf6af-699a-4fa1-b079-3790cd1095e5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.829425] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-881afb25-e805-4aee-8538-bcab26bcfd65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.838603] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 930.838603] env[65385]: value = "task-4453985" [ 930.838603] env[65385]: _type = "Task" [ 930.838603] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.850409] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453985, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.862772] env[65385]: DEBUG nova.compute.manager [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received event network-vif-plugged-2ff7c104-f798-4d67-b92b-3d73b0810a38 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 930.863035] env[65385]: DEBUG oslo_concurrency.lockutils [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Acquiring lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.863256] env[65385]: DEBUG oslo_concurrency.lockutils [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.863402] env[65385]: DEBUG oslo_concurrency.lockutils [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.863571] env[65385]: DEBUG nova.compute.manager [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] No waiting events found dispatching network-vif-plugged-2ff7c104-f798-4d67-b92b-3d73b0810a38 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 930.863806] env[65385]: WARNING nova.compute.manager [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received unexpected event network-vif-plugged-2ff7c104-f798-4d67-b92b-3d73b0810a38 for instance with vm_state active and task_state None. [ 930.864058] env[65385]: DEBUG nova.compute.manager [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received event network-changed-2ff7c104-f798-4d67-b92b-3d73b0810a38 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 930.864248] env[65385]: DEBUG nova.compute.manager [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Refreshing instance network info cache due to event network-changed-2ff7c104-f798-4d67-b92b-3d73b0810a38. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 930.864426] env[65385]: DEBUG oslo_concurrency.lockutils [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Acquiring lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.864566] env[65385]: DEBUG oslo_concurrency.lockutils [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Acquired lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 930.864701] env[65385]: DEBUG nova.network.neutron [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Refreshing network info cache for port 2ff7c104-f798-4d67-b92b-3d73b0810a38 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 930.886868] env[65385]: DEBUG nova.compute.manager [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Received event network-changed-37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 930.887545] env[65385]: DEBUG nova.compute.manager [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Refreshing instance network info cache due to event network-changed-37cdfc1d-4b24-4ae6-ac71-e0d63602b872. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 930.887890] env[65385]: DEBUG oslo_concurrency.lockutils [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Acquiring lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.888273] env[65385]: DEBUG oslo_concurrency.lockutils [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Acquired lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 930.888618] env[65385]: DEBUG nova.network.neutron [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Refreshing network info cache for port 37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 930.993717] env[65385]: DEBUG oslo_vmware.api [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453983, 'name': PowerOnVM_Task, 'duration_secs': 1.105402} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.994054] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 930.994204] env[65385]: INFO nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Took 10.08 seconds to spawn the instance on the hypervisor. [ 930.994380] env[65385]: DEBUG nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 930.995186] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe98b18c-8a79-491c-91cb-30201a5e266b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.234368] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.867s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.237457] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.959s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.237457] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.237537] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 931.237896] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.088s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.239543] env[65385]: INFO nova.compute.claims [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.243734] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35035b08-4047-47d2-aa7f-56d14bf8a754 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.255593] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037a334d-0368-4323-89be-32d5d96630c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.261273] env[65385]: INFO nova.compute.manager [-] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Took 1.69 seconds to deallocate network for instance. [ 931.264257] env[65385]: INFO nova.scheduler.client.report [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Deleted allocations for instance cf37e10a-14f4-4456-bad0-d7528457658d [ 931.280851] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17aa87d9-b54f-4aaf-b076-e58645d006e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.291459] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd58e279-eea2-479a-916b-1287ee20d15b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.332160] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178375MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 931.332403] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.351603] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453985, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.368413] env[65385]: WARNING neutronclient.v2_0.client [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 931.370079] env[65385]: WARNING openstack [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 931.370624] env[65385]: WARNING openstack [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 931.392622] env[65385]: DEBUG nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 931.396671] env[65385]: WARNING neutronclient.v2_0.client [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 931.396671] env[65385]: WARNING openstack [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 931.396671] env[65385]: WARNING openstack [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 931.432557] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 931.432864] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 931.433074] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 931.433304] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 931.433454] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 931.433597] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 931.433799] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.433950] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 931.434132] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 931.434287] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 931.434450] env[65385]: DEBUG nova.virt.hardware [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 931.435417] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9a7b40-28fa-4fa4-9a05-92fe7f28ee66 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.445505] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca12c5df-13c1-4555-b792-59941c9be77c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.464349] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.470813] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Creating folder: Project (aa0f5f9944d64102ac0d26ecade8a54a). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 931.471190] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f991b21-cfa7-4bf8-92cf-9e32cdb5519d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.485948] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Created folder: Project (aa0f5f9944d64102ac0d26ecade8a54a) in parent group-v870881. [ 931.486293] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Creating folder: Instances. Parent ref: group-v871102. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 931.486706] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f133d3a-128c-4b3e-92e4-957d57f2a60c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.501685] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Created folder: Instances in parent group-v871102. [ 931.501804] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 931.502107] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.502453] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-595f7811-5efd-4c1c-ba8b-6f2c679a37b3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.525604] env[65385]: INFO nova.compute.manager [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Took 39.27 seconds to build instance. [ 931.529777] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.529777] env[65385]: value = "task-4453988" [ 931.529777] env[65385]: _type = "Task" [ 931.529777] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.540164] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453988, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.781576] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.791091] env[65385]: DEBUG oslo_concurrency.lockutils [None req-674ffcf6-89b2-4e1c-b090-31c25d343957 tempest-InstanceActionsTestJSON-1192338731 tempest-InstanceActionsTestJSON-1192338731-project-member] Lock "cf37e10a-14f4-4456-bad0-d7528457658d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.251s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.857442] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453985, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.807997} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.859303] env[65385]: WARNING openstack [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 931.860909] env[65385]: WARNING openstack [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 931.867419] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 3febf6af-699a-4fa1-b079-3790cd1095e5/3febf6af-699a-4fa1-b079-3790cd1095e5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.867667] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.868332] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dde694de-481c-4c60-a129-83a2a6bc0bea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.876010] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 931.876010] env[65385]: value = "task-4453989" [ 931.876010] env[65385]: _type = "Task" [ 931.876010] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.885500] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453989, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.952540] env[65385]: WARNING openstack [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 931.952944] env[65385]: WARNING openstack [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.028894] env[65385]: DEBUG oslo_concurrency.lockutils [None req-183bf7cb-7e9a-402a-8683-8db9efb1c320 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.783s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.045662] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453988, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.060257] env[65385]: WARNING neutronclient.v2_0.client [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.060928] env[65385]: WARNING openstack [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.061433] env[65385]: WARNING openstack [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.086836] env[65385]: WARNING neutronclient.v2_0.client [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.087509] env[65385]: WARNING openstack [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.087870] env[65385]: WARNING openstack [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.243453] env[65385]: DEBUG nova.network.neutron [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updated VIF entry in instance network info cache for port 2ff7c104-f798-4d67-b92b-3d73b0810a38. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 932.243885] env[65385]: DEBUG nova.network.neutron [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updating instance_info_cache with network_info: [{"id": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "address": "fa:16:3e:42:06:97", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d96802a-11", "ovs_interfaceid": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2ff7c104-f798-4d67-b92b-3d73b0810a38", "address": "fa:16:3e:e8:96:84", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff7c104-f7", "ovs_interfaceid": "2ff7c104-f798-4d67-b92b-3d73b0810a38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 932.282457] env[65385]: DEBUG nova.network.neutron [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Updated VIF entry in instance network info cache for port 37cdfc1d-4b24-4ae6-ac71-e0d63602b872. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 932.282518] env[65385]: DEBUG nova.network.neutron [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Updating instance_info_cache with network_info: [{"id": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "address": "fa:16:3e:e2:4b:e7", "network": {"id": "8f9a0fa3-6de0-4601-bed3-6ba0670da6e1", "bridge": "br-int", "label": "tempest-ServersTestJSON-1442256052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "02f51c46458e425d9d0c330c02bde184", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0d7a2b2f-3b49-4dc8-9096-af16144b27a9", "external-id": "nsx-vlan-transportzone-492", "segmentation_id": 492, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37cdfc1d-4b", "ovs_interfaceid": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 932.387829] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072261} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.390656] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.392093] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd516a42-e894-4d4c-86c6-621b6997f9af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.420164] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 3febf6af-699a-4fa1-b079-3790cd1095e5/3febf6af-699a-4fa1-b079-3790cd1095e5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.423746] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24ea0ae5-eda0-4087-8abe-241dba7be30d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.445999] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 932.445999] env[65385]: value = "task-4453990" [ 932.445999] env[65385]: _type = "Task" [ 932.445999] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.455986] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453990, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.548773] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4453988, 'name': CreateVM_Task, 'duration_secs': 0.855815} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.549021] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.549875] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.549875] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.550063] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 932.550304] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ab10176-fee0-48da-91e4-919ad8685b3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.562387] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 932.562387] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c9aea6-c6ed-6629-4685-62a0c57610ed" [ 932.562387] env[65385]: _type = "Task" [ 932.562387] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.574574] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c9aea6-c6ed-6629-4685-62a0c57610ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.585977] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c15aded-b8a5-4004-92cb-f5db4972a34b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.596588] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a51b7f6-1297-4b0e-86a3-5c41bdc4303a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.629334] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e78ec9-0175-49b9-8d27-dacc4afebc31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.639291] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2916043e-5c35-4757-b4df-0c0b9ca7e07a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.655536] env[65385]: DEBUG nova.compute.provider_tree [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.750565] env[65385]: DEBUG oslo_concurrency.lockutils [req-a7ae63a5-a9b2-408c-a68a-af831ff55ac3 req-eb0ba52b-e541-4d9c-be6c-ba0ff1a3b891 service nova] Releasing lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 932.791924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.792874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.793208] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "fbc67b68-64ed-410a-a43a-efa3f55b3031-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.793605] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.793864] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.797538] env[65385]: DEBUG oslo_concurrency.lockutils [req-28a463d0-899d-46e5-b54d-64fd2636038b req-6de5670b-274a-4871-9f05-9f009ef726d6 service nova] Releasing lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 932.798884] env[65385]: INFO nova.compute.manager [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Terminating instance [ 932.958732] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453990, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.083201] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c9aea6-c6ed-6629-4685-62a0c57610ed, 'name': SearchDatastore_Task, 'duration_secs': 0.014273} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.083631] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.083906] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.084209] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.084392] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 933.084585] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.084903] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-661ea92d-5c20-4411-aaca-65258d30accb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.096958] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.097172] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 933.098059] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8551ef7-12d0-4caa-b0e9-cbdc4d4644c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.110615] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 933.110615] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf7b87-3342-b46b-5c40-8d61505b92f3" [ 933.110615] env[65385]: _type = "Task" [ 933.110615] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.122322] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf7b87-3342-b46b-5c40-8d61505b92f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.158656] env[65385]: DEBUG nova.scheduler.client.report [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 933.302944] env[65385]: DEBUG nova.compute.manager [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 933.303201] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.304145] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5366e3d3-37b3-4362-a059-1f985d209cbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.315253] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.315553] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-960d51fb-78cf-4c94-990d-25a9cd923f62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.329043] env[65385]: DEBUG nova.compute.manager [req-c86d7ab5-0062-425f-985d-a198ca8885b6 req-21296e3a-9285-4c8b-99fa-aa1ca8cbab51 service nova] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Received event network-vif-deleted-6785f095-0836-4bc7-923a-b16075641a80 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 933.404028] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.404028] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.404028] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleting the datastore file [datastore2] fbc67b68-64ed-410a-a43a-efa3f55b3031 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.404028] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab3a7199-e7da-4275-ade3-7cded490dcfc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.412961] env[65385]: DEBUG oslo_vmware.api [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 933.412961] env[65385]: value = "task-4453992" [ 933.412961] env[65385]: _type = "Task" [ 933.412961] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.422367] env[65385]: DEBUG oslo_vmware.api [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453992, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.458305] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453990, 'name': ReconfigVM_Task, 'duration_secs': 0.692409} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.458613] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 3febf6af-699a-4fa1-b079-3790cd1095e5/3febf6af-699a-4fa1-b079-3790cd1095e5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.459597] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77c84928-4ad6-447b-8b32-511eb4d248e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.470561] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 933.470561] env[65385]: value = "task-4453993" [ 933.470561] env[65385]: _type = "Task" [ 933.470561] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.483318] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453993, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.622954] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf7b87-3342-b46b-5c40-8d61505b92f3, 'name': SearchDatastore_Task, 'duration_secs': 0.011372} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.624667] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b681ebbb-9e70-45da-81fc-995a3a91caf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.631270] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 933.631270] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522744e7-cfd6-d120-4fa4-64c6d117e646" [ 933.631270] env[65385]: _type = "Task" [ 933.631270] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.644035] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522744e7-cfd6-d120-4fa4-64c6d117e646, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.649345] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.649555] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.668294] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.668808] env[65385]: DEBUG nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 933.673685] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.341s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.928047] env[65385]: DEBUG oslo_vmware.api [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4453992, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196363} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.928698] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.929167] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.929167] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.929274] env[65385]: INFO nova.compute.manager [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Took 0.63 seconds to destroy the instance on the hypervisor. [ 933.929554] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 933.929658] env[65385]: DEBUG nova.compute.manager [-] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 933.929753] env[65385]: DEBUG nova.network.neutron [-] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 933.930014] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.930542] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 933.930820] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.981226] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453993, 'name': Rename_Task, 'duration_secs': 0.187467} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.981671] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.981744] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a64c72d-f1ef-4041-8689-dd6cda972087 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.989712] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 933.989712] env[65385]: value = "task-4453994" [ 933.989712] env[65385]: _type = "Task" [ 933.989712] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.998819] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.045757] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.148024] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522744e7-cfd6-d120-4fa4-64c6d117e646, 'name': SearchDatastore_Task, 'duration_secs': 0.014069} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.148024] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 934.148024] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 934.148024] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc0f5974-bdcb-45c9-aabd-91fbbfdeb54a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.152539] env[65385]: DEBUG nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 934.157547] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 934.157547] env[65385]: value = "task-4453995" [ 934.157547] env[65385]: _type = "Task" [ 934.157547] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.168476] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4453995, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.178704] env[65385]: DEBUG nova.compute.utils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 934.188691] env[65385]: DEBUG nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 934.189339] env[65385]: DEBUG nova.network.neutron [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 934.189837] env[65385]: WARNING neutronclient.v2_0.client [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.190485] env[65385]: WARNING neutronclient.v2_0.client [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.191380] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.194036] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.414508] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.414508] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.414508] env[65385]: INFO nova.compute.manager [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Shelving [ 934.470333] env[65385]: DEBUG nova.policy [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c849598ae1e54e37b80ae352d3490fff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b91a3483f4d240de813033cb33d3ee52', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 934.502324] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453994, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.682496] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4453995, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.700244] env[65385]: DEBUG nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 934.715058] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.732645] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 59f26752-75ab-45e5-b0a9-bbcef44ecd7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.732889] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 26a2d464-0aab-463f-8265-948a4f0ff188 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.732979] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0c345794-036c-45d1-985e-cf5c116bd873 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.733091] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance c18b5bde-1a17-4954-a2c5-582815ebfec8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.733336] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0c6206ee-c787-4cd1-9289-1b2620dc4c5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.733336] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 370b6aa8-2337-48e1-8102-611ce4494a55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.733410] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 4b0b8859-b63e-4740-bf67-5733c75e1c70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.733511] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.733653] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance ac6fa516-d954-466e-b8e0-e12440492049 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 934.734400] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance fbc67b68-64ed-410a-a43a-efa3f55b3031 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.734400] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance e8f9418e-904c-4e3f-8ea9-91ab0fdd080d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.734400] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 3febf6af-699a-4fa1-b079-3790cd1095e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.734686] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 354527f0-007d-449f-9e15-48ce1d91876f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.734881] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 934.884574] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.885112] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.885346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.885494] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.889326] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.889326] env[65385]: DEBUG nova.network.neutron [-] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 934.889494] env[65385]: INFO nova.compute.manager [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Terminating instance [ 934.975027] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.975027] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 935.003647] env[65385]: DEBUG oslo_vmware.api [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4453994, 'name': PowerOnVM_Task, 'duration_secs': 0.817233} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.007322] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.007322] env[65385]: INFO nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Took 10.19 seconds to spawn the instance on the hypervisor. [ 935.007322] env[65385]: DEBUG nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 935.007322] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696a370a-6b0f-46e3-af4b-3f7263514e2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.017098] env[65385]: DEBUG nova.network.neutron [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Successfully created port: a986b101-b5ac-4541-93af-25024b4d3cf7 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 935.020573] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.020762] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 935.034658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-c18b5bde-1a17-4954-a2c5-582815ebfec8-2ff7c104-f798-4d67-b92b-3d73b0810a38" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.034987] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-c18b5bde-1a17-4954-a2c5-582815ebfec8-2ff7c104-f798-4d67-b92b-3d73b0810a38" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 935.175853] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4453995, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677282} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.176166] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.176529] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.176663] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb4513ae-11e3-4e9f-a1c7-39d3fe1d519f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.184405] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 935.184405] env[65385]: value = "task-4453996" [ 935.184405] env[65385]: _type = "Task" [ 935.184405] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.193556] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4453996, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.238942] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 935.392436] env[65385]: INFO nova.compute.manager [-] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Took 1.46 seconds to deallocate network for instance. [ 935.397023] env[65385]: DEBUG nova.compute.manager [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 935.397023] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.398576] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c238d2b0-ee88-4c08-8c26-2921b5e9acc3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.408082] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.408619] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3284a59b-3fde-465d-9632-00746f27d49d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.419450] env[65385]: DEBUG oslo_vmware.api [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 935.419450] env[65385]: value = "task-4453997" [ 935.419450] env[65385]: _type = "Task" [ 935.419450] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.428525] env[65385]: DEBUG oslo_vmware.api [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.435737] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.436124] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-300c02de-154b-46f0-9312-2e15b08e990a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.446052] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 935.446052] env[65385]: value = "task-4453998" [ 935.446052] env[65385]: _type = "Task" [ 935.446052] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.460111] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.478304] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 935.531591] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 935.536248] env[65385]: INFO nova.compute.manager [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Took 39.46 seconds to build instance. [ 935.538983] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.539232] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 935.540180] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f04bd8-a6e3-4281-947d-e8bb1706eb7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.561700] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f29e27-fa1f-4c9e-bd85-628f8e5a6601 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.584579] env[65385]: WARNING neutronclient.v2_0.client [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.591373] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Reconfiguring VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 935.592160] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8de0108-7398-4f8a-9d7d-8fc834ea4e3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.617657] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 935.617657] env[65385]: value = "task-4453999" [ 935.617657] env[65385]: _type = "Task" [ 935.617657] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.632218] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.696555] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4453996, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072507} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.697231] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.698026] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f96f9b-3294-4f09-a823-c2c7cab0d5c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.718714] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.720294] env[65385]: DEBUG nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 935.722436] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52508bfd-d26d-47dc-bdb4-9c2b331cd5b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.743361] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 935.746274] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 935.746274] env[65385]: value = "task-4454000" [ 935.746274] env[65385]: _type = "Task" [ 935.746274] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.757073] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454000, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.767111] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 935.767394] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 935.767688] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 935.767841] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 935.768009] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 935.768211] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 935.768446] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.768643] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 935.768929] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 935.769126] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 935.769391] env[65385]: DEBUG nova.virt.hardware [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 935.770389] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1529ff33-98ed-474d-bce0-cea48daefc69 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.780323] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5846a0d-8919-4a12-8ba8-dcf0c726895e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.903389] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.927923] env[65385]: DEBUG oslo_vmware.api [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4453997, 'name': PowerOffVM_Task, 'duration_secs': 0.321615} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.928465] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.928782] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.929214] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fea9b826-84d5-4adf-85fc-75b63d1e5ed5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.955843] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4453998, 'name': PowerOffVM_Task, 'duration_secs': 0.248173} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.956675] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.957621] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08a39d0-d8fd-49e0-b0c1-68d35854d800 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.977414] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318bb3d1-41bf-42fe-b542-527e47b60000 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.012735] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.013625] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.013625] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Deleting the datastore file [datastore2] e8f9418e-904c-4e3f-8ea9-91ab0fdd080d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.013851] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6f152a5-2ce5-4ac8-94f6-b6b3bd14628d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.017181] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.025108] env[65385]: DEBUG oslo_vmware.api [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for the task: (returnval){ [ 936.025108] env[65385]: value = "task-4454002" [ 936.025108] env[65385]: _type = "Task" [ 936.025108] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.034960] env[65385]: DEBUG oslo_vmware.api [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4454002, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.038630] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bd864e01-645a-4f87-9f7e-17c379991492 tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.980s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.053788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.131468] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.251810] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 936.252074] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 936.252526] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=100GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '13', 'num_vm_active': '9', 'num_task_None': '9', 'num_os_type_None': '13', 'num_proj_60de0d1c162342209795a180391661f6': '2', 'io_workload': '3', 'num_proj_2ba61d77cc2d4fe1ba3c03466d5985d9': '2', 'num_task_shelving': '1', 'num_proj_647f4e3157c34a8f9a64cb8ba8b72da6': '1', 'num_proj_1991528775bd4557ae6150b30095e923': '1', 'num_proj_c27eb6232a2148a1a259f57494b4ae30': '1', 'num_proj_b27c092a0fd345ea953956469c21f8bb': '1', 'num_vm_stopped': '1', 'num_task_deleting': '1', 'num_proj_754b1facaaa14501b2204c98e1d7a5a8': '1', 'num_proj_e073e1e0d8a045a48ea8fe3c4dd28b6f': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_02f51c46458e425d9d0c330c02bde184': '1', 'num_proj_aa0f5f9944d64102ac0d26ecade8a54a': '1', 'num_proj_b91a3483f4d240de813033cb33d3ee52': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 936.263593] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454000, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.317285] env[65385]: DEBUG nova.compute.manager [req-a5b5bd81-9b3c-441a-9c96-3cab2d12f7e1 req-2f7ba708-736e-4ce5-9ed5-2f673e0d43b7 service nova] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Received event network-vif-deleted-1a20c5c9-1b05-42cf-a71f-894e2921e893 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 936.494528] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 936.495481] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-07d5ef9f-6f45-4e80-a01d-d7d80325deab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.505738] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 936.505738] env[65385]: value = "task-4454003" [ 936.505738] env[65385]: _type = "Task" [ 936.505738] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.520655] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454003, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.536783] env[65385]: DEBUG oslo_vmware.api [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4454002, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.556843] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e23219-4e94-4be4-bf64-4d11aceff624 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.568287] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80451148-6336-47cf-8c5d-b739a5c2cd76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.609513] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28852e19-85b7-414c-b810-5ee923278696 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.622774] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951f5a00-d941-49b5-a0f7-cb4a0f912d58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.632172] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.640103] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.706737] env[65385]: DEBUG nova.network.neutron [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Successfully updated port: a986b101-b5ac-4541-93af-25024b4d3cf7 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 936.759115] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454000, 'name': ReconfigVM_Task, 'duration_secs': 0.862809} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.759505] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.760248] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8d929768-bccb-43b4-929c-6e2588dbbcdf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.770514] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 936.770514] env[65385]: value = "task-4454004" [ 936.770514] env[65385]: _type = "Task" [ 936.770514] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.780744] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454004, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.018967] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454003, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.035494] env[65385]: DEBUG oslo_vmware.api [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Task: {'id': task-4454002, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.524841} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.035758] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.035970] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 937.036111] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 937.036292] env[65385]: INFO nova.compute.manager [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Took 1.64 seconds to destroy the instance on the hypervisor. [ 937.036516] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 937.036683] env[65385]: DEBUG nova.compute.manager [-] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 937.036805] env[65385]: DEBUG nova.network.neutron [-] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 937.037030] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.037560] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.037812] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 937.129615] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.143925] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 937.212734] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "refresh_cache-78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.212734] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquired lock "refresh_cache-78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 937.212734] env[65385]: DEBUG nova.network.neutron [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 937.282750] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454004, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.422510] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.518650] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454003, 'name': CreateSnapshot_Task, 'duration_secs': 0.843146} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.518852] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 937.519638] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60803fb-dd9d-41dc-b2ef-c8a233518146 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.632955] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.649571] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 937.649571] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.976s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.649745] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.868s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.649946] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.655545] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.940s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.655545] env[65385]: INFO nova.compute.claims [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.690853] env[65385]: INFO nova.scheduler.client.report [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted allocations for instance ac6fa516-d954-466e-b8e0-e12440492049 [ 937.716403] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.716791] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 937.787945] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454004, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.788884] env[65385]: DEBUG nova.network.neutron [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 937.813637] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.814159] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 937.867158] env[65385]: WARNING neutronclient.v2_0.client [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.867866] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.868237] env[65385]: WARNING openstack [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 938.000261] env[65385]: DEBUG nova.network.neutron [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Updating instance_info_cache with network_info: [{"id": "a986b101-b5ac-4541-93af-25024b4d3cf7", "address": "fa:16:3e:c5:29:90", "network": {"id": "570f0ccf-1be7-420f-b9a8-558da905f0c7", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-558987942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b91a3483f4d240de813033cb33d3ee52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa986b101-b5", "ovs_interfaceid": "a986b101-b5ac-4541-93af-25024b4d3cf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 938.040098] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 938.041343] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-991f9dab-b72e-40de-9e8d-90deac3602f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.052032] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 938.052032] env[65385]: value = "task-4454005" [ 938.052032] env[65385]: _type = "Task" [ 938.052032] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.062150] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454005, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.137504] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.202992] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a97b06fd-dc23-4f05-b171-1872dc19c87a tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "ac6fa516-d954-466e-b8e0-e12440492049" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.267s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.259937] env[65385]: DEBUG nova.network.neutron [-] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 938.288130] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454004, 'name': Rename_Task, 'duration_secs': 1.20683} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.288130] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.288130] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2163987b-e6db-47ef-b5b9-7527cebaf437 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.298700] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 938.298700] env[65385]: value = "task-4454006" [ 938.298700] env[65385]: _type = "Task" [ 938.298700] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.309771] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.423733] env[65385]: DEBUG nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Received event network-vif-plugged-a986b101-b5ac-4541-93af-25024b4d3cf7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 938.423939] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Acquiring lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.424956] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.425190] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.425391] env[65385]: DEBUG nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] No waiting events found dispatching network-vif-plugged-a986b101-b5ac-4541-93af-25024b4d3cf7 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 938.426817] env[65385]: WARNING nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Received unexpected event network-vif-plugged-a986b101-b5ac-4541-93af-25024b4d3cf7 for instance with vm_state building and task_state spawning. [ 938.426817] env[65385]: DEBUG nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Received event network-changed-a986b101-b5ac-4541-93af-25024b4d3cf7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 938.426817] env[65385]: DEBUG nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Refreshing instance network info cache due to event network-changed-a986b101-b5ac-4541-93af-25024b4d3cf7. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 938.426817] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Acquiring lock "refresh_cache-78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.503330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Releasing lock "refresh_cache-78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 938.504156] env[65385]: DEBUG nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Instance network_info: |[{"id": "a986b101-b5ac-4541-93af-25024b4d3cf7", "address": "fa:16:3e:c5:29:90", "network": {"id": "570f0ccf-1be7-420f-b9a8-558da905f0c7", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-558987942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b91a3483f4d240de813033cb33d3ee52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa986b101-b5", "ovs_interfaceid": "a986b101-b5ac-4541-93af-25024b4d3cf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 938.504780] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Acquired lock "refresh_cache-78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 938.505211] env[65385]: DEBUG nova.network.neutron [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Refreshing network info cache for port a986b101-b5ac-4541-93af-25024b4d3cf7 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 938.507038] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:29:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a986b101-b5ac-4541-93af-25024b4d3cf7', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.519666] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Creating folder: Project (b91a3483f4d240de813033cb33d3ee52). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.521044] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ab09e2d-ab2e-4ffd-8ea6-7d68ae0efea4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.535349] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Created folder: Project (b91a3483f4d240de813033cb33d3ee52) in parent group-v870881. [ 938.535612] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Creating folder: Instances. Parent ref: group-v871107. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.535909] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e21d8ed7-5c47-4758-b866-31ad0fc2ecaa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.550325] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Created folder: Instances in parent group-v871107. [ 938.550628] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 938.550852] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.554083] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dde69c72-6a01-49fa-9dd0-d03a4df8e954 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.581961] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454005, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.583938] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.583938] env[65385]: value = "task-4454009" [ 938.583938] env[65385]: _type = "Task" [ 938.583938] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.593665] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454009, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.641941] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.763381] env[65385]: INFO nova.compute.manager [-] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Took 1.73 seconds to deallocate network for instance. [ 938.813685] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454006, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.882617] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.882890] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.883144] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.883329] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.883507] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.888316] env[65385]: INFO nova.compute.manager [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Terminating instance [ 938.968700] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacc545b-75a0-4517-86ee-15678a747b12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.979793] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f54081-31d6-4399-af32-152cdbe42479 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.018840] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe9b7c8-0f0a-4e56-b41c-3157f9a372f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.021846] env[65385]: WARNING neutronclient.v2_0.client [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.022625] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.022798] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.037840] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cd45bd-9d7d-44d6-a75b-880ae8cf37cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.054640] env[65385]: DEBUG nova.compute.provider_tree [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.081045] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454005, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.094436] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454009, 'name': CreateVM_Task, 'duration_secs': 0.353356} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.094719] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 939.095200] env[65385]: WARNING neutronclient.v2_0.client [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.095546] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.095716] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.096052] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 939.096325] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51cd9e0a-2950-4203-a456-259f3025069c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.102871] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 939.102871] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5200438a-b11b-d535-e78a-cd990bc5f5e2" [ 939.102871] env[65385]: _type = "Task" [ 939.102871] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.112325] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5200438a-b11b-d535-e78a-cd990bc5f5e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.137066] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.183409] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.183750] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.273448] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.310690] env[65385]: DEBUG oslo_vmware.api [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454006, 'name': PowerOnVM_Task, 'duration_secs': 0.532639} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.311684] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.311684] env[65385]: INFO nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Took 7.92 seconds to spawn the instance on the hypervisor. [ 939.311684] env[65385]: DEBUG nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 939.312759] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4467c05-6d7e-46fb-9049-f0b97956a413 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.317753] env[65385]: WARNING neutronclient.v2_0.client [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.318888] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.319306] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.392926] env[65385]: DEBUG nova.compute.manager [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 939.393178] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.394400] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2bb5c1-53d5-4743-bb83-7c07075b3882 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.406007] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.406007] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-059d9ab2-1b17-48df-aafe-e454f8fb922a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.411696] env[65385]: DEBUG oslo_vmware.api [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 939.411696] env[65385]: value = "task-4454010" [ 939.411696] env[65385]: _type = "Task" [ 939.411696] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.425920] env[65385]: DEBUG oslo_vmware.api [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.558999] env[65385]: DEBUG nova.scheduler.client.report [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 939.587597] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454005, 'name': CloneVM_Task, 'duration_secs': 1.323067} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.590117] env[65385]: DEBUG nova.network.neutron [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Updated VIF entry in instance network info cache for port a986b101-b5ac-4541-93af-25024b4d3cf7. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 939.590499] env[65385]: DEBUG nova.network.neutron [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Updating instance_info_cache with network_info: [{"id": "a986b101-b5ac-4541-93af-25024b4d3cf7", "address": "fa:16:3e:c5:29:90", "network": {"id": "570f0ccf-1be7-420f-b9a8-558da905f0c7", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-558987942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b91a3483f4d240de813033cb33d3ee52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa986b101-b5", "ovs_interfaceid": "a986b101-b5ac-4541-93af-25024b4d3cf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 939.591873] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Created linked-clone VM from snapshot [ 939.595167] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d469663c-f62d-4a60-939f-81fcc3f16a93 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.602252] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Uploading image c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 939.616309] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5200438a-b11b-d535-e78a-cd990bc5f5e2, 'name': SearchDatastore_Task, 'duration_secs': 0.019756} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.616428] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 939.616729] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.617026] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.617187] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.617367] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.617638] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3855fee3-9caa-4e31-9ea4-b798c973e38a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.634432] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.634633] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.635897] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12becdcf-01ba-4e4c-afc1-2c88d8b947fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.642395] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.644489] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 939.644489] env[65385]: value = "vm-871106" [ 939.644489] env[65385]: _type = "VirtualMachine" [ 939.644489] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 939.645089] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1e0dc250-f6a1-4a45-8532-7463bb2cf559 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.648517] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 939.648517] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae4603-f116-bf28-9f28-3c343f7f566b" [ 939.648517] env[65385]: _type = "Task" [ 939.648517] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.654556] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lease: (returnval){ [ 939.654556] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f976ef-f7d7-7bd3-6834-3dc48841507b" [ 939.654556] env[65385]: _type = "HttpNfcLease" [ 939.654556] env[65385]: } obtained for exporting VM: (result){ [ 939.654556] env[65385]: value = "vm-871106" [ 939.654556] env[65385]: _type = "VirtualMachine" [ 939.654556] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 939.654901] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the lease: (returnval){ [ 939.654901] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f976ef-f7d7-7bd3-6834-3dc48841507b" [ 939.654901] env[65385]: _type = "HttpNfcLease" [ 939.654901] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 939.661591] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae4603-f116-bf28-9f28-3c343f7f566b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.665745] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.665745] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f976ef-f7d7-7bd3-6834-3dc48841507b" [ 939.665745] env[65385]: _type = "HttpNfcLease" [ 939.665745] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 939.666038] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 939.666038] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f976ef-f7d7-7bd3-6834-3dc48841507b" [ 939.666038] env[65385]: _type = "HttpNfcLease" [ 939.666038] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 939.666867] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67743210-9e16-4e6e-95a9-d95a47c5616b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.675284] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5224498e-6256-6159-e095-513186c5a180/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 939.675466] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5224498e-6256-6159-e095-513186c5a180/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 939.780838] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f29b28e5-1e65-49f3-9e4e-bd55f2948395 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.852734] env[65385]: INFO nova.compute.manager [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Took 41.77 seconds to build instance. [ 939.855769] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "f4af288e-b9c9-4273-8e65-e726ae039d03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.856018] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.924984] env[65385]: DEBUG oslo_vmware.api [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454010, 'name': PowerOffVM_Task, 'duration_secs': 0.201042} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.925476] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.925735] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.926149] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25b513a0-71d0-4001-a2c7-150ac8c49b85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.997786] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.998498] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.998716] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleting the datastore file [datastore2] 0c6206ee-c787-4cd1-9289-1b2620dc4c5f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.998996] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7ef2741-47a0-4712-b280-739819855a61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.007427] env[65385]: DEBUG oslo_vmware.api [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 940.007427] env[65385]: value = "task-4454013" [ 940.007427] env[65385]: _type = "Task" [ 940.007427] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.018014] env[65385]: DEBUG oslo_vmware.api [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.066019] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.066741] env[65385]: DEBUG nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 940.070048] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.167s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.073801] env[65385]: DEBUG nova.objects.instance [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'resources' on Instance uuid fbc67b68-64ed-410a-a43a-efa3f55b3031 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.094039] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Releasing lock "refresh_cache-78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.094656] env[65385]: DEBUG nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Received event network-changed-37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 940.094874] env[65385]: DEBUG nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Refreshing instance network info cache due to event network-changed-37cdfc1d-4b24-4ae6-ac71-e0d63602b872. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 940.095123] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Acquiring lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.095277] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Acquired lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 940.095450] env[65385]: DEBUG nova.network.neutron [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Refreshing network info cache for port 37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 940.139595] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.162115] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae4603-f116-bf28-9f28-3c343f7f566b, 'name': SearchDatastore_Task, 'duration_secs': 0.018037} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.162973] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bc54cf9-1078-4dd6-b814-f27795a0fc53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.169643] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 940.169643] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cdb780-20c6-9cac-b408-f564c8daa452" [ 940.169643] env[65385]: _type = "Task" [ 940.169643] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.179355] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cdb780-20c6-9cac-b408-f564c8daa452, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.363169] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11d6fbc8-e469-4048-b4dc-9f15b1ea12cc tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "354527f0-007d-449f-9e15-48ce1d91876f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.296s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.363553] env[65385]: DEBUG nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 940.525441] env[65385]: DEBUG oslo_vmware.api [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263013} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.528837] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.528837] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.528837] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.528837] env[65385]: INFO nova.compute.manager [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 940.528837] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 940.529115] env[65385]: DEBUG nova.compute.manager [-] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 940.529184] env[65385]: DEBUG nova.network.neutron [-] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 940.529455] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.530118] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.530417] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.573843] env[65385]: DEBUG nova.compute.utils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 940.577022] env[65385]: DEBUG nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 940.577290] env[65385]: DEBUG nova.network.neutron [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 940.578014] env[65385]: WARNING neutronclient.v2_0.client [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.578499] env[65385]: WARNING neutronclient.v2_0.client [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.579213] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.579717] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.592361] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.599318] env[65385]: WARNING neutronclient.v2_0.client [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.600042] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.600428] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.645302] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.697215] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cdb780-20c6-9cac-b408-f564c8daa452, 'name': SearchDatastore_Task, 'duration_secs': 0.027235} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.703902] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.705056] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b/78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.709461] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60afd259-3bd8-494e-be23-d53b0627432e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.720411] env[65385]: DEBUG nova.policy [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac74a729b62042eda193030d716719e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1598307c304d46a884a88b3eb5f5d7dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 940.734976] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 940.734976] env[65385]: value = "task-4454014" [ 940.734976] env[65385]: _type = "Task" [ 940.734976] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.749411] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454014, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.889994] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.972782] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3c7d9d-ad3a-4129-88e7-4a4c5ad7d8c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.982601] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc978db-2e2c-43b7-98b4-10e4341501a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.022950] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d19eec-302e-4a98-a71b-cee5222330cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.033755] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021e7610-df66-412f-a0ea-11ad517db57e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.051313] env[65385]: DEBUG nova.compute.provider_tree [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.096571] env[65385]: DEBUG nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 941.143661] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.246423] env[65385]: INFO nova.compute.manager [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Rebuilding instance [ 941.257721] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454014, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.319331] env[65385]: DEBUG nova.compute.manager [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 941.320187] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115c7f9f-b833-43c6-8fda-196418c7dc3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.448025] env[65385]: DEBUG nova.network.neutron [-] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 941.556999] env[65385]: DEBUG nova.scheduler.client.report [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 941.641525] env[65385]: DEBUG oslo_vmware.api [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4453999, 'name': ReconfigVM_Task, 'duration_secs': 5.844119} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.643381] env[65385]: DEBUG nova.network.neutron [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Successfully created port: ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 941.646615] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.646856] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Reconfigured VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 941.649387] env[65385]: WARNING neutronclient.v2_0.client [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 941.649862] env[65385]: WARNING neutronclient.v2_0.client [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 941.650884] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.651290] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.722544] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.722544] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.752527] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454014, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.786892} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.752527] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b/78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.753918] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.753918] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20a4f3d9-e71a-4908-b333-121488c4bc3f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.764054] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 941.764054] env[65385]: value = "task-4454015" [ 941.764054] env[65385]: _type = "Task" [ 941.764054] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.780755] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.952270] env[65385]: INFO nova.compute.manager [-] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Took 1.42 seconds to deallocate network for instance. [ 942.063505] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.993s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.067919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.050s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.071123] env[65385]: INFO nova.compute.claims [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.101371] env[65385]: INFO nova.scheduler.client.report [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted allocations for instance fbc67b68-64ed-410a-a43a-efa3f55b3031 [ 942.112156] env[65385]: DEBUG nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 942.154493] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 942.155741] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 942.155741] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 942.155741] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 942.155741] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 942.155741] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 942.155994] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.156064] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 942.156234] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 942.156392] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 942.156560] env[65385]: DEBUG nova.virt.hardware [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 942.158294] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417ecc16-cad4-452d-88cd-a01942d9f9d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.168273] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a822c9-f747-43d0-9423-e5ecfdbab5c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.243164] env[65385]: WARNING neutronclient.v2_0.client [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.243164] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.243164] env[65385]: WARNING openstack [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.277229] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117306} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.277539] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 942.278421] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536d2ab5-5f80-4604-b903-e15a3e4eceb9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.302858] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b/78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.303223] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f79a1940-3491-4f33-9168-b6115f7ed24a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.328270] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 942.328270] env[65385]: value = "task-4454016" [ 942.328270] env[65385]: _type = "Task" [ 942.328270] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.337938] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454016, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.339756] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.340073] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a42214a-c945-4478-9a03-e61922c44474 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.348146] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 942.348146] env[65385]: value = "task-4454017" [ 942.348146] env[65385]: _type = "Task" [ 942.348146] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.361666] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454017, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.464421] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.618325] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2ba6fe6e-f036-470c-b803-e0aba2df5434 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "fbc67b68-64ed-410a-a43a-efa3f55b3031" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.826s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.766206] env[65385]: DEBUG nova.network.neutron [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Updated VIF entry in instance network info cache for port 37cdfc1d-4b24-4ae6-ac71-e0d63602b872. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 942.766762] env[65385]: DEBUG nova.network.neutron [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Updating instance_info_cache with network_info: [{"id": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "address": "fa:16:3e:e2:4b:e7", "network": {"id": "8f9a0fa3-6de0-4601-bed3-6ba0670da6e1", "bridge": "br-int", "label": "tempest-ServersTestJSON-1442256052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "02f51c46458e425d9d0c330c02bde184", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0d7a2b2f-3b49-4dc8-9096-af16144b27a9", "external-id": "nsx-vlan-transportzone-492", "segmentation_id": 492, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37cdfc1d-4b", "ovs_interfaceid": "37cdfc1d-4b24-4ae6-ac71-e0d63602b872", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 942.840741] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.859438] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454017, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.009644] env[65385]: DEBUG nova.compute.manager [req-00170582-3256-4cdc-a995-8acfccff135c req-e933c6fd-7e29-4fff-9722-ccb7b60c4b54 service nova] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Received event network-vif-deleted-d9c28a6e-bfd5-46d3-96bc-ce3090c2a33e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 943.221035] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.221035] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.221035] env[65385]: DEBUG nova.network.neutron [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 943.272449] env[65385]: DEBUG oslo_concurrency.lockutils [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] Releasing lock "refresh_cache-3febf6af-699a-4fa1-b079-3790cd1095e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.272818] env[65385]: DEBUG nova.compute.manager [req-93dba25d-b845-4c95-bd46-127e1077d6c2 req-c1cd04dd-3bff-40fb-9a2d-ed9dd9690b02 service nova] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Received event network-vif-deleted-984014f3-cd87-4054-a971-942d597305f7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 943.343732] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454016, 'name': ReconfigVM_Task, 'duration_secs': 0.676709} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.343989] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b/78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.345721] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4411dc1-aac1-4bd6-a0d4-b0339b9143c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.363332] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 943.363332] env[65385]: value = "task-4454018" [ 943.363332] env[65385]: _type = "Task" [ 943.363332] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.371830] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454017, 'name': PowerOffVM_Task, 'duration_secs': 0.725465} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.379477] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.380291] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 943.380727] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454018, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.384658] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0274545-947b-4eb4-96aa-1c9a2f723abf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.394926] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 943.395309] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df71f92d-b34f-4bbf-94fd-3ee7d94d3408 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.428281] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 943.428561] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 943.428744] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Deleting the datastore file [datastore1] 354527f0-007d-449f-9e15-48ce1d91876f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.429087] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a29322fd-5328-4d80-8d6b-04b9d3cf3f43 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.438146] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 943.438146] env[65385]: value = "task-4454020" [ 943.438146] env[65385]: _type = "Task" [ 943.438146] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.451757] env[65385]: DEBUG nova.compute.manager [req-12a7309d-f889-4af5-929b-b54b76953037 req-38efcce1-d1b7-45a8-ae85-828671c46fd8 service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Received event network-vif-plugged-ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 943.452852] env[65385]: DEBUG oslo_concurrency.lockutils [req-12a7309d-f889-4af5-929b-b54b76953037 req-38efcce1-d1b7-45a8-ae85-828671c46fd8 service nova] Acquiring lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.452852] env[65385]: DEBUG oslo_concurrency.lockutils [req-12a7309d-f889-4af5-929b-b54b76953037 req-38efcce1-d1b7-45a8-ae85-828671c46fd8 service nova] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.452852] env[65385]: DEBUG oslo_concurrency.lockutils [req-12a7309d-f889-4af5-929b-b54b76953037 req-38efcce1-d1b7-45a8-ae85-828671c46fd8 service nova] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.452852] env[65385]: DEBUG nova.compute.manager [req-12a7309d-f889-4af5-929b-b54b76953037 req-38efcce1-d1b7-45a8-ae85-828671c46fd8 service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] No waiting events found dispatching network-vif-plugged-ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 943.452852] env[65385]: WARNING nova.compute.manager [req-12a7309d-f889-4af5-929b-b54b76953037 req-38efcce1-d1b7-45a8-ae85-828671c46fd8 service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Received unexpected event network-vif-plugged-ace973a7-f2da-4905-99cb-e72ce9ef3f18 for instance with vm_state building and task_state spawning. [ 943.458132] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.497154] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2f74db-cef6-4a2f-8ea1-db406bda6d5a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.508743] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc2c7ec-472e-401c-97bc-c8daafb9449c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.550583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3e8099-05e6-4528-be8f-e4a3eae60ce9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.565947] env[65385]: DEBUG nova.network.neutron [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Successfully updated port: ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 943.570987] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf87271-254f-47f2-8266-204fa582ba32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.587126] env[65385]: DEBUG nova.compute.provider_tree [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.719280] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.719280] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.723023] env[65385]: WARNING neutronclient.v2_0.client [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.723775] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.725032] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.875416] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454018, 'name': Rename_Task, 'duration_secs': 0.177377} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.875664] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.876019] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c8f0130-b4f4-4cf7-95df-f5c77d78eaf2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.884923] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 943.884923] env[65385]: value = "task-4454021" [ 943.884923] env[65385]: _type = "Task" [ 943.884923] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.895393] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454021, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.950386] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454020, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121582} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.950658] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.950868] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.951062] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.972876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.973188] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.973413] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.973758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.973876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.978432] env[65385]: INFO nova.compute.manager [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Terminating instance [ 944.035795] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.036205] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.077265] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.077481] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.077665] env[65385]: DEBUG nova.network.neutron [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 944.090169] env[65385]: DEBUG nova.scheduler.client.report [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 944.129580] env[65385]: WARNING neutronclient.v2_0.client [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.130913] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.130913] env[65385]: WARNING openstack [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.221709] env[65385]: DEBUG nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 944.285562] env[65385]: INFO nova.network.neutron [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Port 2ff7c104-f798-4d67-b92b-3d73b0810a38 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 944.285998] env[65385]: DEBUG nova.network.neutron [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updating instance_info_cache with network_info: [{"id": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "address": "fa:16:3e:42:06:97", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d96802a-11", "ovs_interfaceid": "7d96802a-1199-4a18-9a3d-f7b69a52590c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 944.397709] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454021, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.485939] env[65385]: DEBUG nova.compute.manager [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 944.486291] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.488249] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a404f9-f13d-4596-b8b6-f3704daa01b3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.501522] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.501982] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-611898bf-b81e-437b-bf24-c4a695d20496 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.514106] env[65385]: DEBUG oslo_vmware.api [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 944.514106] env[65385]: value = "task-4454022" [ 944.514106] env[65385]: _type = "Task" [ 944.514106] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.532177] env[65385]: DEBUG oslo_vmware.api [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.587771] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.588281] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.597706] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.598812] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 944.603832] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.550s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.606310] env[65385]: INFO nova.compute.claims [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.743969] env[65385]: DEBUG nova.network.neutron [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 944.747756] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.789779] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-c18b5bde-1a17-4954-a2c5-582815ebfec8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.824629] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.825169] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.897265] env[65385]: DEBUG oslo_vmware.api [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454021, 'name': PowerOnVM_Task, 'duration_secs': 0.843708} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.897656] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.897802] env[65385]: INFO nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Took 9.18 seconds to spawn the instance on the hypervisor. [ 944.897981] env[65385]: DEBUG nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 944.898861] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2be62e4-c4ce-4bcf-b681-0f72774d81e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.931912] env[65385]: WARNING neutronclient.v2_0.client [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.932426] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.932803] env[65385]: WARNING openstack [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.994011] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 944.994269] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 944.994437] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 944.994980] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 944.995222] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 944.995386] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 944.995628] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.995864] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 944.996075] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 944.996243] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 944.996412] env[65385]: DEBUG nova.virt.hardware [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 944.997772] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cf2d3c-5ea4-4f2e-abf8-faa8004ce113 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.015428] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d6411e-6f8a-4ef1-a8ce-cdbd9f89533d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.036064] env[65385]: DEBUG oslo_vmware.api [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454022, 'name': PowerOffVM_Task, 'duration_secs': 0.331113} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.046423] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.046785] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.047420] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.055894] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 945.058741] env[65385]: DEBUG nova.network.neutron [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 945.060419] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5694e85-7b4a-47b3-b52d-372b58372580 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.063703] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.063703] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07874be2-e27f-4079-ae6f-81fe39391211 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.080119] env[65385]: DEBUG nova.compute.manager [req-e59ed9b4-9f58-4986-b01a-d8f309a91b1b req-7e60f427-f50f-4b54-a6ef-2f3f0a86830e service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received event network-vif-deleted-2ff7c104-f798-4d67-b92b-3d73b0810a38 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 945.087928] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.087928] env[65385]: value = "task-4454024" [ 945.087928] env[65385]: _type = "Task" [ 945.087928] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.102270] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454024, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.111867] env[65385]: DEBUG nova.compute.utils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 945.119334] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 945.119334] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 945.119334] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.119816] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.120154] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.121275] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.164274] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.164569] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.164777] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleting the datastore file [datastore1] c18b5bde-1a17-4954-a2c5-582815ebfec8 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.165151] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11ad7930-d856-4fc5-87da-865da99c0377 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.174113] env[65385]: DEBUG oslo_vmware.api [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 945.174113] env[65385]: value = "task-4454025" [ 945.174113] env[65385]: _type = "Task" [ 945.174113] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.185611] env[65385]: DEBUG oslo_vmware.api [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.199362] env[65385]: DEBUG nova.policy [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd122bc776500423ab3e717a0f4d5229e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abc72fbdb7e0400b8ce8b8ef0f946fb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 945.202915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.202915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 945.295549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d684c0f6-a6cc-4f19-b79b-2e52e0336e9a tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-c18b5bde-1a17-4954-a2c5-582815ebfec8-2ff7c104-f798-4d67-b92b-3d73b0810a38" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.260s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 945.423757] env[65385]: INFO nova.compute.manager [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Took 22.31 seconds to build instance. [ 945.565349] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.565799] env[65385]: DEBUG nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Instance network_info: |[{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 945.566381] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:f7:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9875d38f-76e2-416c-bfb7-f18a22b0d8ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ace973a7-f2da-4905-99cb-e72ce9ef3f18', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.576989] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 945.578769] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.579463] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2927e426-2a12-4b2a-9557-3a4ed515a07a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.611771] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454024, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.613763] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.613763] env[65385]: value = "task-4454026" [ 945.613763] env[65385]: _type = "Task" [ 945.613763] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.619476] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 945.627779] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454026, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.660236] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Successfully created port: 67be7e47-2dcd-4602-8638-fccd0da10003 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 945.693074] env[65385]: DEBUG oslo_vmware.api [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.381501} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.693676] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.693676] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.693845] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.695185] env[65385]: INFO nova.compute.manager [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Took 1.21 seconds to destroy the instance on the hypervisor. [ 945.695185] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 945.695743] env[65385]: DEBUG nova.compute.manager [-] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 945.695838] env[65385]: DEBUG nova.network.neutron [-] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 945.696971] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.698636] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.699113] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.712021] env[65385]: DEBUG nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 946.629020] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.630708] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ae84c38e-18a3-4665-9b7b-27faae8d799e tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.529s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.641939] env[65385]: DEBUG nova.compute.manager [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Received event network-changed-ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 946.641939] env[65385]: DEBUG nova.compute.manager [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Refreshing instance network info cache due to event network-changed-ace973a7-f2da-4905-99cb-e72ce9ef3f18. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 946.641939] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Acquiring lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.641939] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Acquired lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.641939] env[65385]: DEBUG nova.network.neutron [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Refreshing network info cache for port ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 946.657849] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454024, 'name': CreateVM_Task, 'duration_secs': 1.513346} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.664415] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.664415] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454026, 'name': CreateVM_Task, 'duration_secs': 0.742382} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.664415] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.664415] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.664415] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 946.664415] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.664846] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f15dad25-b6e4-48bd-a368-08914d64b002 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.666865] env[65385]: WARNING neutronclient.v2_0.client [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.667248] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.673228] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.676035] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 946.676035] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ef8b35-d915-33f4-be90-391ee148e9cd" [ 946.676035] env[65385]: _type = "Task" [ 946.676035] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.687928] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ef8b35-d915-33f4-be90-391ee148e9cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.846968] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fba962-b20b-4f2d-a440-95050396a4f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.855248] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2201c85e-cca0-4d0d-9594-8b76a3a05bb4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.886804] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50aeecab-7437-4ac3-a543-94855f9ad32a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.895798] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50db2348-049a-484a-b202-474cc7b968d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.913174] env[65385]: DEBUG nova.compute.provider_tree [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.144944] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 947.150569] env[65385]: WARNING neutronclient.v2_0.client [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.151305] env[65385]: WARNING openstack [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.151775] env[65385]: WARNING openstack [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.178463] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 947.178748] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 947.178910] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 947.179589] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 947.179737] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 947.179939] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 947.180120] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.180306] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 947.181254] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 947.181518] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 947.181753] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 947.182781] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ec7e6a-e0a7-4302-8437-43e88b698448 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.197936] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ef8b35-d915-33f4-be90-391ee148e9cd, 'name': SearchDatastore_Task, 'duration_secs': 0.011442} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.198376] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.198607] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.198851] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.198989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.199173] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.200493] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9301925c-8ddf-4a0f-99c5-adafc4db402c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.206759] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.207133] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 947.207389] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e607eaa-0a25-4692-9dd9-43d90ecf1ed3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.209858] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db108d9e-d30f-4604-87b4-3913dd44a06d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.226115] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 947.226115] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dbc3a5-a871-b4a5-7416-44f832684e62" [ 947.226115] env[65385]: _type = "Task" [ 947.226115] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.226430] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.226571] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.227693] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c32784b1-a54a-4cd6-ab25-1df91677dad2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.238712] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 947.238712] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524207cc-2072-30e1-b17f-6e2eb6e8d476" [ 947.238712] env[65385]: _type = "Task" [ 947.238712] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.242715] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dbc3a5-a871-b4a5-7416-44f832684e62, 'name': SearchDatastore_Task, 'duration_secs': 0.020506} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.246558] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.247402] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.247402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.253744] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524207cc-2072-30e1-b17f-6e2eb6e8d476, 'name': SearchDatastore_Task, 'duration_secs': 0.011913} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.254623] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c28ebd2c-f2e0-42b8-bd92-bbf3367bc6f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.263522] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 947.263522] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525ae509-a233-d4ce-e49b-6f45019ee15c" [ 947.263522] env[65385]: _type = "Task" [ 947.263522] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.274852] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525ae509-a233-d4ce-e49b-6f45019ee15c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.356675] env[65385]: WARNING openstack [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.356675] env[65385]: WARNING openstack [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.416099] env[65385]: DEBUG nova.scheduler.client.report [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 947.453724] env[65385]: WARNING neutronclient.v2_0.client [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.454383] env[65385]: WARNING openstack [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.454719] env[65385]: WARNING openstack [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.500206] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Successfully updated port: 67be7e47-2dcd-4602-8638-fccd0da10003 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 947.600881] env[65385]: DEBUG nova.network.neutron [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updated VIF entry in instance network info cache for port ace973a7-f2da-4905-99cb-e72ce9ef3f18. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 947.600881] env[65385]: DEBUG nova.network.neutron [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.608172] env[65385]: DEBUG nova.network.neutron [-] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.640702] env[65385]: DEBUG nova.compute.manager [req-49fea62e-fc06-4603-905e-f5486e68c7a2 req-7cb11de5-3e72-4362-bddb-e8e0e0af8747 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Received event network-vif-plugged-67be7e47-2dcd-4602-8638-fccd0da10003 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 947.640702] env[65385]: DEBUG oslo_concurrency.lockutils [req-49fea62e-fc06-4603-905e-f5486e68c7a2 req-7cb11de5-3e72-4362-bddb-e8e0e0af8747 service nova] Acquiring lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.640702] env[65385]: DEBUG oslo_concurrency.lockutils [req-49fea62e-fc06-4603-905e-f5486e68c7a2 req-7cb11de5-3e72-4362-bddb-e8e0e0af8747 service nova] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.641018] env[65385]: DEBUG oslo_concurrency.lockutils [req-49fea62e-fc06-4603-905e-f5486e68c7a2 req-7cb11de5-3e72-4362-bddb-e8e0e0af8747 service nova] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.641018] env[65385]: DEBUG nova.compute.manager [req-49fea62e-fc06-4603-905e-f5486e68c7a2 req-7cb11de5-3e72-4362-bddb-e8e0e0af8747 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] No waiting events found dispatching network-vif-plugged-67be7e47-2dcd-4602-8638-fccd0da10003 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 947.641385] env[65385]: WARNING nova.compute.manager [req-49fea62e-fc06-4603-905e-f5486e68c7a2 req-7cb11de5-3e72-4362-bddb-e8e0e0af8747 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Received unexpected event network-vif-plugged-67be7e47-2dcd-4602-8638-fccd0da10003 for instance with vm_state building and task_state spawning. [ 947.778595] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525ae509-a233-d4ce-e49b-6f45019ee15c, 'name': SearchDatastore_Task, 'duration_secs': 0.010666} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.778938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.779171] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.779454] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.779653] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.779907] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7cb8763-0925-44e0-9f0d-5f42253fa678 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.782241] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed74f974-60ad-4bd7-b64b-50d6edff9942 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.790612] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 947.790612] env[65385]: value = "task-4454027" [ 947.790612] env[65385]: _type = "Task" [ 947.790612] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.798531] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.798531] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.799577] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeeed40b-a85c-45fc-93c3-341f89cc0342 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.810382] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454027, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.814538] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 947.814538] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc3ae5-bca6-6e0f-dd4e-38e7272cb243" [ 947.814538] env[65385]: _type = "Task" [ 947.814538] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.827979] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc3ae5-bca6-6e0f-dd4e-38e7272cb243, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.926456] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.323s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.926632] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 947.929270] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.656s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.929551] env[65385]: DEBUG nova.objects.instance [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lazy-loading 'resources' on Instance uuid e8f9418e-904c-4e3f-8ea9-91ab0fdd080d {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.008082] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "refresh_cache-11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.008335] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "refresh_cache-11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 948.008557] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 948.103224] env[65385]: DEBUG oslo_concurrency.lockutils [req-ae733321-b781-45f7-b97b-fca56f34259d req-7799effc-d49d-423f-87b3-a92a99ed095a service nova] Releasing lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.113325] env[65385]: INFO nova.compute.manager [-] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Took 2.42 seconds to deallocate network for instance. [ 948.306408] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454027, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512033} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.306746] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.306961] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.307267] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ac4c634-401d-49e8-8aef-90d711173eda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.316238] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 948.316238] env[65385]: value = "task-4454028" [ 948.316238] env[65385]: _type = "Task" [ 948.316238] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.331580] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc3ae5-bca6-6e0f-dd4e-38e7272cb243, 'name': SearchDatastore_Task, 'duration_secs': 0.010522} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.337183] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454028, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.337183] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64d5f6ff-3653-4c12-8c35-127edb5cecff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.345904] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 948.345904] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5208d571-6563-7026-219d-5d075de56446" [ 948.345904] env[65385]: _type = "Task" [ 948.345904] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.358317] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5208d571-6563-7026-219d-5d075de56446, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.433276] env[65385]: DEBUG nova.compute.utils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 948.439593] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 948.440825] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 948.441523] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.442050] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.442665] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.443271] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.482470] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5224498e-6256-6159-e095-513186c5a180/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 948.483802] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1de93a4-ea90-421c-9043-fc26ea952d56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.491502] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5224498e-6256-6159-e095-513186c5a180/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 948.491911] env[65385]: ERROR oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5224498e-6256-6159-e095-513186c5a180/disk-0.vmdk due to incomplete transfer. [ 948.492190] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f2c48d53-183f-4473-a97d-ef6ceac233ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.505215] env[65385]: DEBUG oslo_vmware.rw_handles [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5224498e-6256-6159-e095-513186c5a180/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 948.506132] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Uploaded image c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 948.510546] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 948.517803] env[65385]: DEBUG nova.policy [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd122bc776500423ab3e717a0f4d5229e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abc72fbdb7e0400b8ce8b8ef0f946fb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 948.519859] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3ff5dc3f-da16-4c06-800b-e88f05c7fc43 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.523439] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.523824] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.542295] env[65385]: DEBUG nova.compute.manager [req-89c87419-14fb-42c7-8665-c400d06f200b req-0f9b3d6b-677e-49ff-b02e-bbf659af257b service nova] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Received event network-vif-deleted-7d96802a-1199-4a18-9a3d-f7b69a52590c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 948.543380] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "167e1b46-af52-4a7a-9964-edabc8155287" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.546073] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "167e1b46-af52-4a7a-9964-edabc8155287" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.549960] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 948.549960] env[65385]: value = "task-4454029" [ 948.549960] env[65385]: _type = "Task" [ 948.549960] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.569565] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454029, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.612708] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 948.621503] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.662298] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.662682] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.762263] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.763052] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.763417] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.819146] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf014ee-186e-4252-b30f-837baa3c4f25 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.834840] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934e2d14-f215-497d-a83b-9fd0be490071 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.838849] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454028, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143129} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.839551] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.841085] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dba094-8f17-4429-89b2-6dd9fb9405f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.871664] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Successfully created port: 978b2387-8e83-4565-9a81-983a73480cf1 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 948.876253] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Updating instance_info_cache with network_info: [{"id": "67be7e47-2dcd-4602-8638-fccd0da10003", "address": "fa:16:3e:e7:2d:35", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67be7e47-2d", "ovs_interfaceid": "67be7e47-2dcd-4602-8638-fccd0da10003", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 948.882472] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6828decc-2a74-4acf-a778-2aa6231e13b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.903132] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.905115] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bb2560a-b310-4afe-bd93-a6445b02efad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.927736] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5208d571-6563-7026-219d-5d075de56446, 'name': SearchDatastore_Task, 'duration_secs': 0.017223} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.929549] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa87f0c-13f0-4822-84b6-4f61bee543e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.934090] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.934406] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.936299] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d41ab866-89e6-4944-8aa1-c6d8f9c22dc1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.940288] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 948.940288] env[65385]: value = "task-4454030" [ 948.940288] env[65385]: _type = "Task" [ 948.940288] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.941533] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 948.958139] env[65385]: DEBUG nova.compute.provider_tree [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.963355] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 948.963355] env[65385]: value = "task-4454031" [ 948.963355] env[65385]: _type = "Task" [ 948.963355] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.969357] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.977955] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454031, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.050508] env[65385]: DEBUG nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 949.063639] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454029, 'name': Destroy_Task, 'duration_secs': 0.40657} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.063940] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Destroyed the VM [ 949.064232] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 949.064844] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3362fd59-3a92-4e9a-89c1-a5075ec791da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.074862] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 949.074862] env[65385]: value = "task-4454032" [ 949.074862] env[65385]: _type = "Task" [ 949.074862] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.085562] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454032, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.378986] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "refresh_cache-11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.379502] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Instance network_info: |[{"id": "67be7e47-2dcd-4602-8638-fccd0da10003", "address": "fa:16:3e:e7:2d:35", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67be7e47-2d", "ovs_interfaceid": "67be7e47-2dcd-4602-8638-fccd0da10003", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 949.379987] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:2d:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '67be7e47-2dcd-4602-8638-fccd0da10003', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.388113] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Creating folder: Project (abc72fbdb7e0400b8ce8b8ef0f946fb5). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.388397] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8efe3463-f263-4ab6-b0a6-6824e28ba890 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.405103] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Created folder: Project (abc72fbdb7e0400b8ce8b8ef0f946fb5) in parent group-v870881. [ 949.405103] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Creating folder: Instances. Parent ref: group-v871112. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.405103] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac3875a8-66de-4a30-907e-c537101222ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.417979] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Created folder: Instances in parent group-v871112. [ 949.418308] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 949.418522] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.418743] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa379a76-011a-41a6-817b-29d6eb972b56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.440511] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.440511] env[65385]: value = "task-4454035" [ 949.440511] env[65385]: _type = "Task" [ 949.440511] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.455983] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454035, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.469731] env[65385]: DEBUG nova.scheduler.client.report [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 949.473521] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.488981] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454031, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513551} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.489614] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 949.489861] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.490245] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64666d2a-d62d-4ae5-ae87-15c2dd605e8d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.498962] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 949.498962] env[65385]: value = "task-4454036" [ 949.498962] env[65385]: _type = "Task" [ 949.498962] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.512248] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.581237] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.586440] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454032, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.672080] env[65385]: DEBUG nova.compute.manager [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Received event network-changed-67be7e47-2dcd-4602-8638-fccd0da10003 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 949.672080] env[65385]: DEBUG nova.compute.manager [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Refreshing instance network info cache due to event network-changed-67be7e47-2dcd-4602-8638-fccd0da10003. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 949.672497] env[65385]: DEBUG oslo_concurrency.lockutils [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Acquiring lock "refresh_cache-11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.672497] env[65385]: DEBUG oslo_concurrency.lockutils [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Acquired lock "refresh_cache-11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.672497] env[65385]: DEBUG nova.network.neutron [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Refreshing network info cache for port 67be7e47-2dcd-4602-8638-fccd0da10003 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 949.957585] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454035, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.964271] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 949.966982] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454030, 'name': ReconfigVM_Task, 'duration_secs': 0.618076} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.967625] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 354527f0-007d-449f-9e15-48ce1d91876f/354527f0-007d-449f-9e15-48ce1d91876f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.968507] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-917e57c9-7c7e-433e-94ba-6125db7690fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.976210] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.047s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.981156] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.091s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.984134] env[65385]: INFO nova.compute.claims [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.988089] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 949.988089] env[65385]: value = "task-4454037" [ 949.988089] env[65385]: _type = "Task" [ 949.988089] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.999703] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 949.999987] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 950.000349] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 950.000447] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 950.000576] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 950.000694] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 950.000926] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.001134] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 950.001313] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 950.001653] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 950.001768] env[65385]: DEBUG nova.virt.hardware [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 950.003468] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7366eb69-bc44-4d08-a7cd-dd75d65c18e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.013509] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454037, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.014707] env[65385]: INFO nova.scheduler.client.report [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Deleted allocations for instance e8f9418e-904c-4e3f-8ea9-91ab0fdd080d [ 950.026718] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc962497-7d41-4f2b-a93a-a49ab845f0cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.030941] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.134249} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.031267] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.032919] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-500e9681-8881-473f-b24d-9ed172419a96 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.068661] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.069312] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cc457a4-3ea4-4574-ae73-734afb7cf794 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.095800] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454032, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.098375] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 950.098375] env[65385]: value = "task-4454038" [ 950.098375] env[65385]: _type = "Task" [ 950.098375] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.107305] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454038, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.177298] env[65385]: WARNING neutronclient.v2_0.client [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.178034] env[65385]: WARNING openstack [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.178401] env[65385]: WARNING openstack [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.186492] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 950.187072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.187072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 950.187758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.187758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.189701] env[65385]: INFO nova.compute.manager [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Terminating instance [ 950.457392] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454035, 'name': CreateVM_Task, 'duration_secs': 0.828394} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.457568] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.458090] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.458393] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.458536] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.458833] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 950.459102] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7589cb08-046e-4c58-99a8-da16c6ceb5d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.464157] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 950.464157] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d7e6f-15f4-a29b-01a0-36bd9ef09240" [ 950.464157] env[65385]: _type = "Task" [ 950.464157] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.473590] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d7e6f-15f4-a29b-01a0-36bd9ef09240, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.496171] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Successfully updated port: 978b2387-8e83-4565-9a81-983a73480cf1 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 950.509849] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454037, 'name': Rename_Task, 'duration_secs': 0.211665} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.511085] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.511399] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d4827655-50bc-4cc5-bff2-5fd04d0c5690 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.521905] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 950.521905] env[65385]: value = "task-4454039" [ 950.521905] env[65385]: _type = "Task" [ 950.521905] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.530783] env[65385]: DEBUG oslo_concurrency.lockutils [None req-edb295ec-2b8b-4096-bc1f-70705caabae7 tempest-ServerPasswordTestJSON-1118711915 tempest-ServerPasswordTestJSON-1118711915-project-member] Lock "e8f9418e-904c-4e3f-8ea9-91ab0fdd080d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.643s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.535400] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.597722] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454032, 'name': RemoveSnapshot_Task, 'duration_secs': 1.321813} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.598030] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 950.598307] env[65385]: DEBUG nova.compute.manager [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 950.601892] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ec0fa7-8ff4-4ca4-ba43-00a7b09de21a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.610513] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454038, 'name': ReconfigVM_Task, 'duration_secs': 0.344187} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.612105] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.615014] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0ed46f2-e9e7-4dd9-b6e7-df5ee1fd2a86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.622712] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 950.622712] env[65385]: value = "task-4454040" [ 950.622712] env[65385]: _type = "Task" [ 950.622712] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.632391] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454040, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.694885] env[65385]: DEBUG nova.compute.manager [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 950.695177] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 950.696475] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f40171-18e8-45f6-a937-4eb191814f72 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.709633] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.710074] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01c9869a-af5f-4def-b4c7-567d6f522122 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.718777] env[65385]: DEBUG oslo_vmware.api [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 950.718777] env[65385]: value = "task-4454041" [ 950.718777] env[65385]: _type = "Task" [ 950.718777] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.730475] env[65385]: DEBUG oslo_vmware.api [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.736405] env[65385]: WARNING openstack [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.736799] env[65385]: WARNING openstack [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.825287] env[65385]: WARNING neutronclient.v2_0.client [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.825946] env[65385]: WARNING openstack [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.826316] env[65385]: WARNING openstack [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.921596] env[65385]: DEBUG nova.network.neutron [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Updated VIF entry in instance network info cache for port 67be7e47-2dcd-4602-8638-fccd0da10003. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 950.922096] env[65385]: DEBUG nova.network.neutron [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Updating instance_info_cache with network_info: [{"id": "67be7e47-2dcd-4602-8638-fccd0da10003", "address": "fa:16:3e:e7:2d:35", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67be7e47-2d", "ovs_interfaceid": "67be7e47-2dcd-4602-8638-fccd0da10003", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 950.978832] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d7e6f-15f4-a29b-01a0-36bd9ef09240, 'name': SearchDatastore_Task, 'duration_secs': 0.010577} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.979296] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.979553] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.979805] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.979951] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.980149] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.980496] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff5a6241-b114-4d26-aa72-e19017b879b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.997049] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.997294] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.998548] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "refresh_cache-f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.998548] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "refresh_cache-f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.998720] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 951.000080] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30113543-b1b5-4aba-8701-3bf7a85e5010 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.011447] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 951.011447] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521a02af-5629-2186-158d-95b969a42ea3" [ 951.011447] env[65385]: _type = "Task" [ 951.011447] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.023313] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521a02af-5629-2186-158d-95b969a42ea3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.034286] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454039, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.122686] env[65385]: INFO nova.compute.manager [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Shelve offloading [ 951.141875] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454040, 'name': Rename_Task, 'duration_secs': 0.214799} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.141875] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.141875] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bb38012-5d69-4ae8-bf10-827a13a8cbaf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.152826] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 951.152826] env[65385]: value = "task-4454042" [ 951.152826] env[65385]: _type = "Task" [ 951.152826] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.168074] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.237641] env[65385]: DEBUG oslo_vmware.api [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454041, 'name': PowerOffVM_Task, 'duration_secs': 0.35444} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.238233] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.238651] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.243532] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-853f6a88-066f-46fe-a750-f55032da79b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.314300] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.314503] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.314676] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Deleting the datastore file [datastore2] 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.314952] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c3e0784-b394-4f4b-8ebe-e72bae5e28eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.326602] env[65385]: DEBUG oslo_vmware.api [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for the task: (returnval){ [ 951.326602] env[65385]: value = "task-4454044" [ 951.326602] env[65385]: _type = "Task" [ 951.326602] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.338993] env[65385]: DEBUG oslo_vmware.api [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454044, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.370796] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963e0dfc-ad7c-4343-a94b-a818f5390bce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.379903] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa322e0-869e-4dff-9158-cf70a5b3b4d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.413277] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c842d366-52cd-43c6-bb0c-693c39fe616f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.422060] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f733a4f-781b-4657-9c0b-33e7ff5785a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.428180] env[65385]: DEBUG oslo_concurrency.lockutils [req-b2ad45ac-479f-4c9b-9124-f87c86dc08c5 req-d283ecba-c157-4d7e-9b37-47b5e08dcd88 service nova] Releasing lock "refresh_cache-11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.439588] env[65385]: DEBUG nova.compute.provider_tree [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.504852] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.505338] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.524820] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521a02af-5629-2186-158d-95b969a42ea3, 'name': SearchDatastore_Task, 'duration_secs': 0.010433} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.529790] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ce8b148-8e0e-44fb-a00f-5def03e5d52f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.539167] env[65385]: DEBUG oslo_vmware.api [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454039, 'name': PowerOnVM_Task, 'duration_secs': 0.721554} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.539305] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.539502] env[65385]: DEBUG nova.compute.manager [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 951.539954] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 951.539954] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf4ea1-99b8-9cd4-7c86-8b8b41169a6f" [ 951.539954] env[65385]: _type = "Task" [ 951.539954] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.540773] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c57ecdb-c802-4742-90f8-45f89370baa6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.558515] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cf4ea1-99b8-9cd4-7c86-8b8b41169a6f, 'name': SearchDatastore_Task, 'duration_secs': 0.013455} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.558967] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.560269] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3/11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.560269] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fef826a6-2497-4bfe-877e-dc8c2db9ef5b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.569644] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 951.569644] env[65385]: value = "task-4454045" [ 951.569644] env[65385]: _type = "Task" [ 951.569644] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.591765] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.622871] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 951.638153] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.638750] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2175967-238c-4df5-b503-b2fe900ae209 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.649408] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 951.649408] env[65385]: value = "task-4454046" [ 951.649408] env[65385]: _type = "Task" [ 951.649408] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.668180] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.668180] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.683289] env[65385]: DEBUG oslo_vmware.api [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454042, 'name': PowerOnVM_Task, 'duration_secs': 0.514745} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.684439] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 951.684439] env[65385]: DEBUG nova.compute.manager [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 951.684439] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.684439] env[65385]: INFO nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Took 9.57 seconds to spawn the instance on the hypervisor. [ 951.684587] env[65385]: DEBUG nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 951.685332] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbe8fca-fc65-484f-85ba-ce1b2605526f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.690936] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb41401-5309-41e4-bf56-1be90fd6d4bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.706274] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.707093] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.707093] env[65385]: DEBUG nova.network.neutron [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 951.768695] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.769581] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.770016] env[65385]: WARNING openstack [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.807658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "4b0b8859-b63e-4740-bf67-5733c75e1c70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.807658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.807658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "4b0b8859-b63e-4740-bf67-5733c75e1c70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.807658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.807658] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.808870] env[65385]: INFO nova.compute.manager [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Terminating instance [ 951.847099] env[65385]: DEBUG oslo_vmware.api [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Task: {'id': task-4454044, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131036} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.847546] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.847546] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.847698] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.847858] env[65385]: INFO nova.compute.manager [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 951.848165] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 951.849097] env[65385]: DEBUG nova.compute.manager [-] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 951.849097] env[65385]: DEBUG nova.network.neutron [-] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 951.849097] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.850041] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.850289] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.879960] env[65385]: DEBUG nova.compute.manager [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Received event network-vif-plugged-978b2387-8e83-4565-9a81-983a73480cf1 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 951.880222] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Acquiring lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.880429] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.880594] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.880758] env[65385]: DEBUG nova.compute.manager [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] No waiting events found dispatching network-vif-plugged-978b2387-8e83-4565-9a81-983a73480cf1 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 951.880964] env[65385]: WARNING nova.compute.manager [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Received unexpected event network-vif-plugged-978b2387-8e83-4565-9a81-983a73480cf1 for instance with vm_state building and task_state spawning. [ 951.881154] env[65385]: DEBUG nova.compute.manager [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Received event network-changed-978b2387-8e83-4565-9a81-983a73480cf1 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 951.881303] env[65385]: DEBUG nova.compute.manager [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Refreshing instance network info cache due to event network-changed-978b2387-8e83-4565-9a81-983a73480cf1. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 951.881485] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Acquiring lock "refresh_cache-f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.889033] env[65385]: DEBUG nova.network.neutron [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Updating instance_info_cache with network_info: [{"id": "978b2387-8e83-4565-9a81-983a73480cf1", "address": "fa:16:3e:36:70:60", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap978b2387-8e", "ovs_interfaceid": "978b2387-8e83-4565-9a81-983a73480cf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.918364] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.944603] env[65385]: DEBUG nova.scheduler.client.report [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 952.070380] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.084750] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454045, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.284104] env[65385]: WARNING neutronclient.v2_0.client [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.284104] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.284104] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.284104] env[65385]: INFO nova.compute.manager [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Took 17.57 seconds to build instance. [ 952.314855] env[65385]: DEBUG nova.compute.manager [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 952.315197] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.316444] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7599bbea-0423-4cfa-9f80-a0fb66968b1c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.337620] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.338265] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88df38c3-9c16-42e7-9595-c38856b1aaad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.347181] env[65385]: DEBUG oslo_vmware.api [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 952.347181] env[65385]: value = "task-4454047" [ 952.347181] env[65385]: _type = "Task" [ 952.347181] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.359479] env[65385]: DEBUG oslo_vmware.api [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454047, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.361403] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.361924] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.393179] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "refresh_cache-f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.393565] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Instance network_info: |[{"id": "978b2387-8e83-4565-9a81-983a73480cf1", "address": "fa:16:3e:36:70:60", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap978b2387-8e", "ovs_interfaceid": "978b2387-8e83-4565-9a81-983a73480cf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 952.394705] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Acquired lock "refresh_cache-f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.394705] env[65385]: DEBUG nova.network.neutron [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Refreshing network info cache for port 978b2387-8e83-4565-9a81-983a73480cf1 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 952.397147] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:70:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '978b2387-8e83-4565-9a81-983a73480cf1', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.405453] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 952.412142] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.415212] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9de0062-0da4-40b3-9f27-2006c7cc0393 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.437590] env[65385]: DEBUG nova.compute.manager [req-2bbd60a7-c0e6-4feb-869c-66766a9161b3 req-606762bd-e1ee-4fce-8f31-4c61a3dacb7c service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Received event network-vif-deleted-a986b101-b5ac-4541-93af-25024b4d3cf7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 952.437797] env[65385]: INFO nova.compute.manager [req-2bbd60a7-c0e6-4feb-869c-66766a9161b3 req-606762bd-e1ee-4fce-8f31-4c61a3dacb7c service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Neutron deleted interface a986b101-b5ac-4541-93af-25024b4d3cf7; detaching it from the instance and deleting it from the info cache [ 952.437856] env[65385]: DEBUG nova.network.neutron [req-2bbd60a7-c0e6-4feb-869c-66766a9161b3 req-606762bd-e1ee-4fce-8f31-4c61a3dacb7c service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.453157] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.453667] env[65385]: DEBUG nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 952.456612] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.456612] env[65385]: value = "task-4454048" [ 952.456612] env[65385]: _type = "Task" [ 952.456612] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.458018] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.994s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.458956] env[65385]: DEBUG nova.objects.instance [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lazy-loading 'resources' on Instance uuid 0c6206ee-c787-4cd1-9289-1b2620dc4c5f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.475689] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454048, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.482086] env[65385]: WARNING neutronclient.v2_0.client [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.482940] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.483377] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.584349] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539609} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.584691] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3/11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.584939] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.585276] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13ddad66-05d9-45c3-a32e-952e6dd40b06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.594912] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 952.594912] env[65385]: value = "task-4454049" [ 952.594912] env[65385]: _type = "Task" [ 952.594912] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.607388] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454049, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.627604] env[65385]: DEBUG nova.network.neutron [-] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.655689] env[65385]: DEBUG nova.network.neutron [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbadc19c-78", "ovs_interfaceid": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.736271] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c875bda9-77a4-42df-98d8-5d556956a174 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.086s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.779344] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.779485] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.779735] env[65385]: INFO nova.compute.manager [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Shelving [ 952.864496] env[65385]: DEBUG oslo_vmware.api [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454047, 'name': PowerOffVM_Task, 'duration_secs': 0.38624} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.864787] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 952.864965] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 952.865877] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-955c5170-9ee8-49c7-ba38-184a334c47d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.900038] env[65385]: WARNING neutronclient.v2_0.client [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.900038] env[65385]: WARNING openstack [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.900038] env[65385]: WARNING openstack [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.942043] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-49e8468e-326d-40d0-924e-8bddc35c1c81 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.957074] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12dec76e-7d74-4e29-9761-e9cc785842a1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.971258] env[65385]: DEBUG nova.compute.utils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 952.982547] env[65385]: DEBUG nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 952.982547] env[65385]: DEBUG nova.network.neutron [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 952.982999] env[65385]: WARNING neutronclient.v2_0.client [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.983618] env[65385]: WARNING neutronclient.v2_0.client [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.985125] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.985125] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.997912] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 952.998259] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 952.999608] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleting the datastore file [datastore1] 4b0b8859-b63e-4740-bf67-5733c75e1c70 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.999608] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a730bf14-d94d-41b0-9cef-71e70946ddf9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.007861] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454048, 'name': CreateVM_Task, 'duration_secs': 0.510115} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.022909] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.023855] env[65385]: DEBUG nova.compute.manager [req-2bbd60a7-c0e6-4feb-869c-66766a9161b3 req-606762bd-e1ee-4fce-8f31-4c61a3dacb7c service nova] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Detach interface failed, port_id=a986b101-b5ac-4541-93af-25024b4d3cf7, reason: Instance 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 953.031124] env[65385]: WARNING neutronclient.v2_0.client [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.031124] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.032024] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.032024] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 953.036258] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-941d294f-afb4-44d2-9105-cf56934631e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.036258] env[65385]: DEBUG oslo_vmware.api [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 953.036258] env[65385]: value = "task-4454051" [ 953.036258] env[65385]: _type = "Task" [ 953.036258] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.044316] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 953.044316] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a1062b-b6d7-52a4-874a-70db040ea800" [ 953.044316] env[65385]: _type = "Task" [ 953.044316] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.052026] env[65385]: DEBUG oslo_vmware.api [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454051, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.058903] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a1062b-b6d7-52a4-874a-70db040ea800, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.070092] env[65385]: DEBUG nova.policy [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de0f069272314a31ad33c77137513a17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c27eb6232a2148a1a259f57494b4ae30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 953.086345] env[65385]: WARNING openstack [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.086785] env[65385]: WARNING openstack [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.106915] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454049, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125474} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.107218] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 953.108036] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a73c61-cf9e-4653-b1f7-e8ab45b9ccc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.133160] env[65385]: INFO nova.compute.manager [-] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Took 1.28 seconds to deallocate network for instance. [ 953.141826] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3/11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.147919] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b92f2a7f-337c-44b4-bad2-16554555748b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.175439] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.175838] env[65385]: WARNING neutronclient.v2_0.client [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.176558] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.177153] env[65385]: WARNING openstack [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.183037] env[65385]: WARNING neutronclient.v2_0.client [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.193310] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 953.193310] env[65385]: value = "task-4454052" [ 953.193310] env[65385]: _type = "Task" [ 953.193310] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.207490] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.392222] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c3b735-191e-4446-92b9-43bea3f4d830 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.402206] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0571d43-c786-4d14-9689-a987e2aea56e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.439661] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "354527f0-007d-449f-9e15-48ce1d91876f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.439949] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "354527f0-007d-449f-9e15-48ce1d91876f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.440207] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "354527f0-007d-449f-9e15-48ce1d91876f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.440355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "354527f0-007d-449f-9e15-48ce1d91876f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.440516] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "354527f0-007d-449f-9e15-48ce1d91876f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.443189] env[65385]: INFO nova.compute.manager [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Terminating instance [ 953.449862] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876ddea3-1436-4746-a74d-5107e32163b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.461078] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2af328-04be-4a44-b246-974a9cd70711 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.480407] env[65385]: DEBUG nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 953.485037] env[65385]: DEBUG nova.compute.provider_tree [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.489951] env[65385]: DEBUG nova.network.neutron [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Successfully created port: 2a55996b-b9ef-4e79-9979-f42e6bdd2127 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 953.552735] env[65385]: DEBUG oslo_vmware.api [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454051, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360163} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.554327] env[65385]: WARNING neutronclient.v2_0.client [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.555322] env[65385]: WARNING openstack [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.555718] env[65385]: WARNING openstack [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.568168] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.568386] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.568597] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.568792] env[65385]: INFO nova.compute.manager [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Took 1.25 seconds to destroy the instance on the hypervisor. [ 953.569052] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 953.569905] env[65385]: DEBUG nova.compute.manager [-] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 953.570017] env[65385]: DEBUG nova.network.neutron [-] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 953.570254] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.570834] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.571209] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.584393] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a1062b-b6d7-52a4-874a-70db040ea800, 'name': SearchDatastore_Task, 'duration_secs': 0.033151} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.584895] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.585034] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.585270] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.585424] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.585595] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.585889] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e4ed6df-d2ce-4b37-80e7-42a9a88f1a7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.598931] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.599129] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.599971] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8def1341-07be-4a0c-ac32-b9b018a6f261 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.606249] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 953.606249] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5221a134-039c-3801-05f9-ad50c527f819" [ 953.606249] env[65385]: _type = "Task" [ 953.606249] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.615574] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5221a134-039c-3801-05f9-ad50c527f819, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.675542] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.706868] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454052, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.720847] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.794526] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.794932] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b7afb47-0e24-4adb-b3c4-1c74bc4238ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.804070] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 953.804070] env[65385]: value = "task-4454053" [ 953.804070] env[65385]: _type = "Task" [ 953.804070] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.814493] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.888029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.888029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.944522] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.945715] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce661185-39b0-43cf-87d0-f45b44ec075b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.955872] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "refresh_cache-354527f0-007d-449f-9e15-48ce1d91876f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.956037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquired lock "refresh_cache-354527f0-007d-449f-9e15-48ce1d91876f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.956176] env[65385]: DEBUG nova.network.neutron [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 953.957597] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.958426] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-135fef39-0cd5-4358-9597-4c0c805f9c2c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.989879] env[65385]: DEBUG nova.network.neutron [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Updated VIF entry in instance network info cache for port 978b2387-8e83-4565-9a81-983a73480cf1. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 953.990353] env[65385]: DEBUG nova.network.neutron [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Updating instance_info_cache with network_info: [{"id": "978b2387-8e83-4565-9a81-983a73480cf1", "address": "fa:16:3e:36:70:60", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap978b2387-8e", "ovs_interfaceid": "978b2387-8e83-4565-9a81-983a73480cf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 953.992788] env[65385]: DEBUG nova.scheduler.client.report [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.042378] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.043157] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.043641] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleting the datastore file [datastore2] 0c345794-036c-45d1-985e-cf5c116bd873 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.044117] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb577728-0f34-4479-bb4d-2da2193077c5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.057313] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 954.057313] env[65385]: value = "task-4454055" [ 954.057313] env[65385]: _type = "Task" [ 954.057313] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.068636] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454055, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.122081] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5221a134-039c-3801-05f9-ad50c527f819, 'name': SearchDatastore_Task, 'duration_secs': 0.033689} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.122081] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-786dfab6-c67c-4d6e-af69-20b2183dea77 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.128916] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 954.128916] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527c9ac8-0d01-9975-ba47-2d21b0cd4ebb" [ 954.128916] env[65385]: _type = "Task" [ 954.128916] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.139687] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527c9ac8-0d01-9975-ba47-2d21b0cd4ebb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.210567] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454052, 'name': ReconfigVM_Task, 'duration_secs': 0.910524} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.210739] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3/11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.211388] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67c61fe8-912b-4631-9373-35abce209e17 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.219750] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 954.219750] env[65385]: value = "task-4454056" [ 954.219750] env[65385]: _type = "Task" [ 954.219750] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.230440] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454056, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.318028] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454053, 'name': PowerOffVM_Task, 'duration_secs': 0.228968} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.318028] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.318028] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551c1cd8-6785-4174-a676-a0759804fe4c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.338195] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c306c1cd-1ab2-48d1-a412-ff1bb4a08355 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.398489] env[65385]: DEBUG nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 954.462023] env[65385]: WARNING neutronclient.v2_0.client [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.462023] env[65385]: WARNING openstack [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.462023] env[65385]: WARNING openstack [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.503619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.045s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.508052] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf75a761-ba0b-4b82-8eb8-de5d1d3b50f9 req-4e6be578-d1bf-41c5-8803-f842b0b9ea26 service nova] Releasing lock "refresh_cache-f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.513360] env[65385]: DEBUG nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 954.513979] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.766s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.516034] env[65385]: INFO nova.compute.claims [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.556766] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 954.557164] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 954.557606] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 954.557606] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 954.557751] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 954.557855] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 954.558115] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.558295] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 954.558472] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 954.558637] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 954.558813] env[65385]: DEBUG nova.virt.hardware [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 954.559997] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05997b3-a532-4970-a640-6a1d02379e30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.564973] env[65385]: INFO nova.scheduler.client.report [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted allocations for instance 0c6206ee-c787-4cd1-9289-1b2620dc4c5f [ 954.580144] env[65385]: DEBUG oslo_vmware.api [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454055, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177719} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.583142] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.583142] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.583322] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.586936] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97e0076-bcfd-4fbc-a642-e97a3316fd30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.606111] env[65385]: DEBUG nova.network.neutron [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 954.619743] env[65385]: INFO nova.scheduler.client.report [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleted allocations for instance 0c345794-036c-45d1-985e-cf5c116bd873 [ 954.641162] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527c9ac8-0d01-9975-ba47-2d21b0cd4ebb, 'name': SearchDatastore_Task, 'duration_secs': 0.010642} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.644255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.644255] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5/f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.644870] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc6ec406-7caa-4dba-903a-9f01c896d99d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.653385] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 954.653385] env[65385]: value = "task-4454057" [ 954.653385] env[65385]: _type = "Task" [ 954.653385] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.663759] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.734616] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454056, 'name': Rename_Task, 'duration_secs': 0.168798} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.735047] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.735267] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe96feaa-b86b-4890-a0e0-b9221c6ee626 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.743878] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 954.743878] env[65385]: value = "task-4454058" [ 954.743878] env[65385]: _type = "Task" [ 954.743878] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.756836] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.761084] env[65385]: DEBUG nova.network.neutron [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 954.844486] env[65385]: DEBUG nova.compute.manager [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Received event network-changed-ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 954.844738] env[65385]: DEBUG nova.compute.manager [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Refreshing instance network info cache due to event network-changed-ace973a7-f2da-4905-99cb-e72ce9ef3f18. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 954.844895] env[65385]: DEBUG oslo_concurrency.lockutils [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Acquiring lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.845059] env[65385]: DEBUG oslo_concurrency.lockutils [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Acquired lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.845633] env[65385]: DEBUG nova.network.neutron [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Refreshing network info cache for port ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 954.850902] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 954.851627] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-37677320-be8d-4ad8-95f8-11b9a5038fc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.862673] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 954.862673] env[65385]: value = "task-4454059" [ 954.862673] env[65385]: _type = "Task" [ 954.862673] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.875867] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454059, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.926203] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.028268] env[65385]: DEBUG nova.network.neutron [-] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 955.077938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ccbab55e-2219-4369-a9fa-374db2a7c2b5 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "0c6206ee-c787-4cd1-9289-1b2620dc4c5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.195s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.125133] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.166986] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454057, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.222864] env[65385]: DEBUG nova.compute.manager [req-90ee065d-ec8b-402e-ab0d-f8b6386cc296 req-3a2a2112-d0cd-451d-ad02-ec1a87982250 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Received event network-vif-plugged-2a55996b-b9ef-4e79-9979-f42e6bdd2127 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 955.222951] env[65385]: DEBUG oslo_concurrency.lockutils [req-90ee065d-ec8b-402e-ab0d-f8b6386cc296 req-3a2a2112-d0cd-451d-ad02-ec1a87982250 service nova] Acquiring lock "f4af288e-b9c9-4273-8e65-e726ae039d03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.223249] env[65385]: DEBUG oslo_concurrency.lockutils [req-90ee065d-ec8b-402e-ab0d-f8b6386cc296 req-3a2a2112-d0cd-451d-ad02-ec1a87982250 service nova] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.223476] env[65385]: DEBUG oslo_concurrency.lockutils [req-90ee065d-ec8b-402e-ab0d-f8b6386cc296 req-3a2a2112-d0cd-451d-ad02-ec1a87982250 service nova] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.223660] env[65385]: DEBUG nova.compute.manager [req-90ee065d-ec8b-402e-ab0d-f8b6386cc296 req-3a2a2112-d0cd-451d-ad02-ec1a87982250 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] No waiting events found dispatching network-vif-plugged-2a55996b-b9ef-4e79-9979-f42e6bdd2127 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 955.224062] env[65385]: WARNING nova.compute.manager [req-90ee065d-ec8b-402e-ab0d-f8b6386cc296 req-3a2a2112-d0cd-451d-ad02-ec1a87982250 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Received unexpected event network-vif-plugged-2a55996b-b9ef-4e79-9979-f42e6bdd2127 for instance with vm_state building and task_state spawning. [ 955.248849] env[65385]: DEBUG nova.network.neutron [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Successfully updated port: 2a55996b-b9ef-4e79-9979-f42e6bdd2127 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 955.258469] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454058, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.263397] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Releasing lock "refresh_cache-354527f0-007d-449f-9e15-48ce1d91876f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.263803] env[65385]: DEBUG nova.compute.manager [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 955.264063] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.264905] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849ea51f-9a43-4420-b05e-408eb0007290 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.275135] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.275396] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cc44f65-5fe4-462c-bdc8-76ab9502dcdf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.282683] env[65385]: DEBUG oslo_vmware.api [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 955.282683] env[65385]: value = "task-4454060" [ 955.282683] env[65385]: _type = "Task" [ 955.282683] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.292343] env[65385]: DEBUG oslo_vmware.api [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.348347] env[65385]: WARNING neutronclient.v2_0.client [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.349010] env[65385]: WARNING openstack [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.349390] env[65385]: WARNING openstack [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.374373] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454059, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.470058] env[65385]: WARNING openstack [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.470479] env[65385]: WARNING openstack [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.533924] env[65385]: INFO nova.compute.manager [-] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Took 1.96 seconds to deallocate network for instance. [ 955.552974] env[65385]: WARNING neutronclient.v2_0.client [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.554430] env[65385]: WARNING openstack [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.554430] env[65385]: WARNING openstack [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.669973] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454057, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548339} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.672731] env[65385]: DEBUG nova.network.neutron [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updated VIF entry in instance network info cache for port ace973a7-f2da-4905-99cb-e72ce9ef3f18. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 955.672853] env[65385]: DEBUG nova.network.neutron [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 955.674262] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5/f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.674525] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.675733] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34a9e9a0-2b66-4ede-9726-e5cc5c65daec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.686839] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 955.686839] env[65385]: value = "task-4454061" [ 955.686839] env[65385]: _type = "Task" [ 955.686839] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.701115] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454061, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.751898] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "refresh_cache-f4af288e-b9c9-4273-8e65-e726ae039d03" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.752128] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "refresh_cache-f4af288e-b9c9-4273-8e65-e726ae039d03" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 955.752375] env[65385]: DEBUG nova.network.neutron [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 955.762070] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454058, 'name': PowerOnVM_Task, 'duration_secs': 0.943661} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.763072] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.763279] env[65385]: INFO nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Took 8.62 seconds to spawn the instance on the hypervisor. [ 955.763455] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 955.764283] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfe779c-cf22-4db4-b3c4-143d0b639e4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.798108] env[65385]: DEBUG oslo_vmware.api [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454060, 'name': PowerOffVM_Task, 'duration_secs': 0.304666} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.798388] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.798569] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.798827] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd1d30d0-453a-4e48-bf94-593153ea35aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.831783] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.831783] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.831783] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Deleting the datastore file [datastore2] 354527f0-007d-449f-9e15-48ce1d91876f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.831783] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69968d46-a835-44e8-8f96-a26d06ee7c40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.835394] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84123925-7a4b-44d8-8a8a-4dff36fc63c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.846230] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4a51a5-6664-441e-bbd8-ca984fa8c861 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.849826] env[65385]: DEBUG oslo_vmware.api [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for the task: (returnval){ [ 955.849826] env[65385]: value = "task-4454063" [ 955.849826] env[65385]: _type = "Task" [ 955.849826] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.887338] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d78a6f-e305-4611-ab0c-09e8b1efb8b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.895056] env[65385]: DEBUG oslo_vmware.api [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454063, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.903263] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee9d3b7-da03-45be-8216-081943312dcf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.907985] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454059, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.921366] env[65385]: DEBUG nova.compute.provider_tree [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.042413] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.180734] env[65385]: DEBUG oslo_concurrency.lockutils [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Releasing lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 956.181033] env[65385]: DEBUG nova.compute.manager [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received event network-vif-unplugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 956.181223] env[65385]: DEBUG oslo_concurrency.lockutils [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.181425] env[65385]: DEBUG oslo_concurrency.lockutils [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.181583] env[65385]: DEBUG oslo_concurrency.lockutils [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.181803] env[65385]: DEBUG nova.compute.manager [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] No waiting events found dispatching network-vif-unplugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 956.181985] env[65385]: WARNING nova.compute.manager [req-a56c1570-0521-4709-b520-4ca42d2e0702 req-70ee9246-2fc3-4730-9870-7043045a608b service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received unexpected event network-vif-unplugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 for instance with vm_state shelved_offloaded and task_state None. [ 956.198514] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454061, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073046} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.198786] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.199621] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33004dc7-0072-4656-8d22-6b0c4dd740d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.223296] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5/f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.223625] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ce8e0b8-4cf8-4b26-a5ad-67e1608dbed1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.245091] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 956.245091] env[65385]: value = "task-4454064" [ 956.245091] env[65385]: _type = "Task" [ 956.245091] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.253969] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454064, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.258623] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.259072] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.289998] env[65385]: INFO nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Took 20.30 seconds to build instance. [ 956.362397] env[65385]: DEBUG oslo_vmware.api [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Task: {'id': task-4454063, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111573} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.363562] env[65385]: DEBUG nova.network.neutron [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 956.365924] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.366877] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.367487] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.371550] env[65385]: INFO nova.compute.manager [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 956.371550] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 956.371550] env[65385]: DEBUG nova.compute.manager [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 956.371550] env[65385]: DEBUG nova.network.neutron [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 956.371550] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.371550] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.371550] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.393194] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454059, 'name': CreateSnapshot_Task, 'duration_secs': 1.1101} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.393481] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 956.394340] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800df458-26f6-4302-a77e-1af7ac1ad0d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.415232] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.415232] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.425782] env[65385]: DEBUG nova.scheduler.client.report [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 956.430449] env[65385]: DEBUG nova.network.neutron [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 956.432018] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.527781] env[65385]: WARNING neutronclient.v2_0.client [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.528472] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.528816] env[65385]: WARNING openstack [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.667333] env[65385]: DEBUG nova.network.neutron [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Updating instance_info_cache with network_info: [{"id": "2a55996b-b9ef-4e79-9979-f42e6bdd2127", "address": "fa:16:3e:49:a8:ad", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a55996b-b9", "ovs_interfaceid": "2a55996b-b9ef-4e79-9979-f42e6bdd2127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.757900] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454064, 'name': ReconfigVM_Task, 'duration_secs': 0.31193} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.760110] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Reconfigured VM instance instance-00000052 to attach disk [datastore1] f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5/f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.763681] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-848f2646-8ae7-4cdc-a79b-d5be42dd6256 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.770891] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 956.770891] env[65385]: value = "task-4454065" [ 956.770891] env[65385]: _type = "Task" [ 956.770891] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.781220] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454065, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.793367] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.820s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.877887] env[65385]: DEBUG nova.compute.manager [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received event network-changed-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 956.878152] env[65385]: DEBUG nova.compute.manager [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Refreshing instance network info cache due to event network-changed-bbadc19c-78ee-4246-a53d-77f01c5ccea8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 956.878307] env[65385]: DEBUG oslo_concurrency.lockutils [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Acquiring lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.878440] env[65385]: DEBUG oslo_concurrency.lockutils [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Acquired lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 956.878813] env[65385]: DEBUG nova.network.neutron [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Refreshing network info cache for port bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 956.938078] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 956.938955] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.939496] env[65385]: DEBUG nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 956.942173] env[65385]: DEBUG nova.network.neutron [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.943822] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-48776e64-6b7a-46d9-834a-cb24fe23ec58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.947285] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.275s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.948728] env[65385]: INFO nova.compute.claims [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.958518] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 956.958518] env[65385]: value = "task-4454066" [ 956.958518] env[65385]: _type = "Task" [ 956.958518] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.968230] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454066, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.171283] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "refresh_cache-f4af288e-b9c9-4273-8e65-e726ae039d03" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 957.171601] env[65385]: DEBUG nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Instance network_info: |[{"id": "2a55996b-b9ef-4e79-9979-f42e6bdd2127", "address": "fa:16:3e:49:a8:ad", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a55996b-b9", "ovs_interfaceid": "2a55996b-b9ef-4e79-9979-f42e6bdd2127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 957.172394] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:a8:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a55996b-b9ef-4e79-9979-f42e6bdd2127', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.184195] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 957.185030] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.185030] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de59cbb2-33fd-4fba-823c-881a5ca96592 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.214682] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.214682] env[65385]: value = "task-4454067" [ 957.214682] env[65385]: _type = "Task" [ 957.214682] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.226667] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454067, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.256533] env[65385]: DEBUG nova.compute.manager [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Received event network-changed-2a55996b-b9ef-4e79-9979-f42e6bdd2127 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 957.256533] env[65385]: DEBUG nova.compute.manager [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Refreshing instance network info cache due to event network-changed-2a55996b-b9ef-4e79-9979-f42e6bdd2127. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 957.256533] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Acquiring lock "refresh_cache-f4af288e-b9c9-4273-8e65-e726ae039d03" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.256871] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Acquired lock "refresh_cache-f4af288e-b9c9-4273-8e65-e726ae039d03" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.256871] env[65385]: DEBUG nova.network.neutron [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Refreshing network info cache for port 2a55996b-b9ef-4e79-9979-f42e6bdd2127 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 957.261047] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.283687] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454065, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.382245] env[65385]: WARNING neutronclient.v2_0.client [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.383032] env[65385]: WARNING openstack [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.383364] env[65385]: WARNING openstack [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.448165] env[65385]: INFO nova.compute.manager [-] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Took 1.08 seconds to deallocate network for instance. [ 957.449679] env[65385]: DEBUG nova.compute.utils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 957.452590] env[65385]: DEBUG nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 957.453082] env[65385]: DEBUG nova.network.neutron [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 957.453158] env[65385]: WARNING neutronclient.v2_0.client [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.453430] env[65385]: WARNING neutronclient.v2_0.client [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.454657] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.455249] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.485705] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454066, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.519988] env[65385]: DEBUG nova.policy [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a710ed03c944a5f830a99d80b4d5ce2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ab35da093ff491fa13580ad2b1e610f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 957.524171] env[65385]: WARNING openstack [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.524736] env[65385]: WARNING openstack [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.620783] env[65385]: WARNING neutronclient.v2_0.client [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.621627] env[65385]: WARNING openstack [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.622095] env[65385]: WARNING openstack [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.728583] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454067, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.750208] env[65385]: DEBUG nova.network.neutron [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updated VIF entry in instance network info cache for port bbadc19c-78ee-4246-a53d-77f01c5ccea8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 957.750624] env[65385]: DEBUG nova.network.neutron [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": null, "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapbbadc19c-78", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.760648] env[65385]: WARNING neutronclient.v2_0.client [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.761474] env[65385]: WARNING openstack [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.761911] env[65385]: WARNING openstack [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.783292] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454065, 'name': Rename_Task, 'duration_secs': 0.904045} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.783682] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.783963] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d807bf3-e729-4476-baf7-70bb79315989 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.792664] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 957.792664] env[65385]: value = "task-4454068" [ 957.792664] env[65385]: _type = "Task" [ 957.792664] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.805752] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454068, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.889294] env[65385]: WARNING openstack [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.889775] env[65385]: WARNING openstack [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.911417] env[65385]: DEBUG nova.network.neutron [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Successfully created port: 8d177373-bd8d-475c-81b6-972d8eae4f95 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 957.967369] env[65385]: DEBUG nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 957.975220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.990425] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454066, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.997575] env[65385]: WARNING neutronclient.v2_0.client [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.998332] env[65385]: WARNING openstack [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.998715] env[65385]: WARNING openstack [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.094441] env[65385]: DEBUG nova.network.neutron [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Updated VIF entry in instance network info cache for port 2a55996b-b9ef-4e79-9979-f42e6bdd2127. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 958.094605] env[65385]: DEBUG nova.network.neutron [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Updating instance_info_cache with network_info: [{"id": "2a55996b-b9ef-4e79-9979-f42e6bdd2127", "address": "fa:16:3e:49:a8:ad", "network": {"id": "429d0fa5-2ae4-4339-81f1-85d2f96da9c0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1393191600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "c27eb6232a2148a1a259f57494b4ae30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a55996b-b9", "ovs_interfaceid": "2a55996b-b9ef-4e79-9979-f42e6bdd2127", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 958.231579] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454067, 'name': CreateVM_Task, 'duration_secs': 0.636538} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.231951] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.232659] env[65385]: WARNING neutronclient.v2_0.client [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.233215] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.233463] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.233911] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 958.238792] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0376884-9e57-4f5d-8169-26826af46f15 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.247450] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 958.247450] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f7c7bd-3d90-9d99-aa8e-a411fe536f48" [ 958.247450] env[65385]: _type = "Task" [ 958.247450] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.253722] env[65385]: DEBUG oslo_concurrency.lockutils [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] Releasing lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.254032] env[65385]: DEBUG nova.compute.manager [req-b0739b04-f167-49d0-8101-2026938c79d3 req-ee81a2e1-a92c-4e3c-92bd-0f4425d79f62 service nova] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Received event network-vif-deleted-2bb8c7e3-459c-4149-8960-0d0df350dc4a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 958.262389] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f7c7bd-3d90-9d99-aa8e-a411fe536f48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.285959] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63499db-86bd-4245-ad54-f3919b782bc4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.299964] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c0c33e-e326-4adf-bcb3-a173d43b9de7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.311657] env[65385]: DEBUG oslo_vmware.api [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454068, 'name': PowerOnVM_Task, 'duration_secs': 0.518758} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.340762] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.341035] env[65385]: INFO nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Took 8.38 seconds to spawn the instance on the hypervisor. [ 958.341221] env[65385]: DEBUG nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 958.342408] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464ad2fc-cddd-43c3-904f-8c3bae1d6b7d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.346033] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fc6791-982c-4de8-b0af-102d62577ea2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.355943] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c291873-4999-411d-bfd7-cff52fd59d7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.374466] env[65385]: DEBUG nova.compute.provider_tree [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.488501] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454066, 'name': CloneVM_Task, 'duration_secs': 1.206593} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.488683] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Created linked-clone VM from snapshot [ 958.489451] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38695bc6-1877-41aa-a582-562679f68071 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.498156] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Uploading image 1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 958.527103] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 958.527103] env[65385]: value = "vm-871117" [ 958.527103] env[65385]: _type = "VirtualMachine" [ 958.527103] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 958.527588] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0b7e0e51-fceb-4638-b8fb-27fde50a6975 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.541442] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease: (returnval){ [ 958.541442] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261eea8-fde7-c1bd-b218-343af49f9d38" [ 958.541442] env[65385]: _type = "HttpNfcLease" [ 958.541442] env[65385]: } obtained for exporting VM: (result){ [ 958.541442] env[65385]: value = "vm-871117" [ 958.541442] env[65385]: _type = "VirtualMachine" [ 958.541442] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 958.541992] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the lease: (returnval){ [ 958.541992] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261eea8-fde7-c1bd-b218-343af49f9d38" [ 958.541992] env[65385]: _type = "HttpNfcLease" [ 958.541992] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 958.549944] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.549944] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261eea8-fde7-c1bd-b218-343af49f9d38" [ 958.549944] env[65385]: _type = "HttpNfcLease" [ 958.549944] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.599320] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b52b87e-a813-42e6-a8af-050b41e8f8b9 req-3b5fd6b3-b49d-455a-8b8f-851db35061a7 service nova] Releasing lock "refresh_cache-f4af288e-b9c9-4273-8e65-e726ae039d03" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.758500] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f7c7bd-3d90-9d99-aa8e-a411fe536f48, 'name': SearchDatastore_Task, 'duration_secs': 0.023246} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.758500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.758746] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.759061] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.759228] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.759411] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.759698] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-091695ae-8790-4a3d-91a3-61046e1d5567 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.775638] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.775860] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.776734] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52f90f82-4b38-4174-87e4-12da314c41e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.783737] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 958.783737] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526b84e9-7b37-1810-7da8-04d71f74f2af" [ 958.783737] env[65385]: _type = "Task" [ 958.783737] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.795575] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526b84e9-7b37-1810-7da8-04d71f74f2af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.870138] env[65385]: INFO nova.compute.manager [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Took 22.83 seconds to build instance. [ 958.879052] env[65385]: DEBUG nova.scheduler.client.report [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 958.986079] env[65385]: DEBUG nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 959.015672] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 959.015918] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 959.016082] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 959.016280] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 959.016508] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 959.016690] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 959.017127] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.017357] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 959.017541] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 959.017722] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 959.017901] env[65385]: DEBUG nova.virt.hardware [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 959.018857] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c037b21e-873b-4eb8-b6ee-c629d49dfe95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.028839] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542457be-2a80-4f82-a0d4-82c9865daa45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.053585] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.053585] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261eea8-fde7-c1bd-b218-343af49f9d38" [ 959.053585] env[65385]: _type = "HttpNfcLease" [ 959.053585] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 959.055346] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 959.055346] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261eea8-fde7-c1bd-b218-343af49f9d38" [ 959.055346] env[65385]: _type = "HttpNfcLease" [ 959.055346] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 959.055346] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b60cf8-f0d4-41d5-904f-91ab8522a4b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.063203] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ed37a-7507-25de-b4f6-d6b081440980/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 959.063372] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ed37a-7507-25de-b4f6-d6b081440980/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 959.172478] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2db33786-5266-4c47-a6ea-094ed60293a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.300820] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526b84e9-7b37-1810-7da8-04d71f74f2af, 'name': SearchDatastore_Task, 'duration_secs': 0.021303} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.301747] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a785f3f-a090-4e9e-ac33-78f8bf77bc91 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.308520] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 959.308520] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f5b16-878a-29a0-d88c-9430a5e86aff" [ 959.308520] env[65385]: _type = "Task" [ 959.308520] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.323397] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f5b16-878a-29a0-d88c-9430a5e86aff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.372853] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0281c389-f82e-4b31-b4bb-653c9189e637 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.352s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.385206] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.385687] env[65385]: DEBUG nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 959.388447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.767s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.388704] env[65385]: DEBUG nova.objects.instance [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'resources' on Instance uuid c18b5bde-1a17-4954-a2c5-582815ebfec8 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.474313] env[65385]: DEBUG nova.network.neutron [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Successfully updated port: 8d177373-bd8d-475c-81b6-972d8eae4f95 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 959.498437] env[65385]: DEBUG nova.compute.manager [req-006d1568-aa0d-489a-938c-177da9312a32 req-d709ead0-4767-4c66-8410-d1ded4910db9 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Received event network-vif-plugged-8d177373-bd8d-475c-81b6-972d8eae4f95 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 959.498659] env[65385]: DEBUG oslo_concurrency.lockutils [req-006d1568-aa0d-489a-938c-177da9312a32 req-d709ead0-4767-4c66-8410-d1ded4910db9 service nova] Acquiring lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.498868] env[65385]: DEBUG oslo_concurrency.lockutils [req-006d1568-aa0d-489a-938c-177da9312a32 req-d709ead0-4767-4c66-8410-d1ded4910db9 service nova] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.499314] env[65385]: DEBUG oslo_concurrency.lockutils [req-006d1568-aa0d-489a-938c-177da9312a32 req-d709ead0-4767-4c66-8410-d1ded4910db9 service nova] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.499314] env[65385]: DEBUG nova.compute.manager [req-006d1568-aa0d-489a-938c-177da9312a32 req-d709ead0-4767-4c66-8410-d1ded4910db9 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] No waiting events found dispatching network-vif-plugged-8d177373-bd8d-475c-81b6-972d8eae4f95 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 959.499574] env[65385]: WARNING nova.compute.manager [req-006d1568-aa0d-489a-938c-177da9312a32 req-d709ead0-4767-4c66-8410-d1ded4910db9 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Received unexpected event network-vif-plugged-8d177373-bd8d-475c-81b6-972d8eae4f95 for instance with vm_state building and task_state spawning. [ 959.718768] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.719045] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.719930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.719930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.719930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.722647] env[65385]: INFO nova.compute.manager [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Terminating instance [ 959.797342] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.797615] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.798253] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.798253] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.798372] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.800585] env[65385]: INFO nova.compute.manager [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Terminating instance [ 959.821503] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f5b16-878a-29a0-d88c-9430a5e86aff, 'name': SearchDatastore_Task, 'duration_secs': 0.029027} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.822380] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.823441] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f4af288e-b9c9-4273-8e65-e726ae039d03/f4af288e-b9c9-4273-8e65-e726ae039d03.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.823441] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-324da7dd-2088-44f3-9e61-df55f3d84ced {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.832717] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 959.832717] env[65385]: value = "task-4454070" [ 959.832717] env[65385]: _type = "Task" [ 959.832717] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.842678] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.898181] env[65385]: DEBUG nova.compute.utils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 959.898181] env[65385]: DEBUG nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 959.898536] env[65385]: DEBUG nova.network.neutron [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 959.898925] env[65385]: WARNING neutronclient.v2_0.client [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.899251] env[65385]: WARNING neutronclient.v2_0.client [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.899908] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.900275] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.959822] env[65385]: DEBUG nova.policy [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d5103a5769a4296973351194f49fed3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '754b1facaaa14501b2204c98e1d7a5a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 959.979764] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "refresh_cache-55936da6-fcdf-4291-9b3a-23fa8b7a56ef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.980138] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquired lock "refresh_cache-55936da6-fcdf-4291-9b3a-23fa8b7a56ef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 959.980377] env[65385]: DEBUG nova.network.neutron [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 960.228759] env[65385]: DEBUG nova.compute.manager [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 960.229666] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.230561] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d57865-18ab-425b-9ff5-e466081e75c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.241256] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.242647] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-367b4f89-6b4a-4879-b29d-3e0b78d0ca50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.245246] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e8766e-69a1-4651-b5f0-a2a15afc757e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.257544] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2cc5e2-cca4-43b5-892f-be4a76066402 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.262441] env[65385]: DEBUG oslo_vmware.api [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 960.262441] env[65385]: value = "task-4454071" [ 960.262441] env[65385]: _type = "Task" [ 960.262441] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.303766] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577494c6-30bf-4c94-a8eb-12c1d8624faf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.312230] env[65385]: DEBUG nova.compute.manager [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 960.312332] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.312637] env[65385]: DEBUG oslo_vmware.api [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.313505] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156eada7-9db1-45e4-b9e8-a05bb16fe99a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.324609] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.328465] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83fd3b86-2195-42e6-940c-cf21bf9cb23c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.331981] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41207699-f430-4ebb-90cc-320c08f39108 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.344811] env[65385]: DEBUG oslo_vmware.api [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 960.344811] env[65385]: value = "task-4454072" [ 960.344811] env[65385]: _type = "Task" [ 960.344811] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.355013] env[65385]: DEBUG nova.compute.provider_tree [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.366631] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454070, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.367257] env[65385]: DEBUG nova.network.neutron [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Successfully created port: 82a01c37-b04c-4b2c-a613-dc4117b64ef2 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 960.380476] env[65385]: DEBUG oslo_vmware.api [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.412474] env[65385]: DEBUG nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 960.483440] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.483977] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.538459] env[65385]: DEBUG nova.network.neutron [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 960.634802] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.635697] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.777419] env[65385]: WARNING neutronclient.v2_0.client [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.778215] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.778372] env[65385]: WARNING openstack [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.796408] env[65385]: DEBUG oslo_vmware.api [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454071, 'name': PowerOffVM_Task, 'duration_secs': 0.374687} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.797907] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.798283] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 960.798604] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d0d9299-c143-4ba9-a646-382db7e656da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.848591] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454070, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.665419} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.848895] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f4af288e-b9c9-4273-8e65-e726ae039d03/f4af288e-b9c9-4273-8e65-e726ae039d03.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.849125] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.849421] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5a27705-47f2-4391-87a7-afe58b64149e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.860086] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 960.860086] env[65385]: value = "task-4454074" [ 960.860086] env[65385]: _type = "Task" [ 960.860086] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.867936] env[65385]: DEBUG nova.scheduler.client.report [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 960.871786] env[65385]: DEBUG oslo_vmware.api [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454072, 'name': PowerOffVM_Task, 'duration_secs': 0.263253} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.872776] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.872953] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 960.873415] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b70427a-4d0b-4434-97b9-889b8a118fbe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.881528] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454074, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.881711] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 960.882019] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 960.882232] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleting the datastore file [datastore1] f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.882935] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-101e7649-0768-4b17-b55e-2fa1baccfd66 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.894892] env[65385]: DEBUG oslo_vmware.api [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 960.894892] env[65385]: value = "task-4454076" [ 960.894892] env[65385]: _type = "Task" [ 960.894892] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.902526] env[65385]: DEBUG oslo_vmware.api [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.976024] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 960.976024] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 960.976024] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleting the datastore file [datastore2] 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.976024] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d845fb35-0fb7-40d7-9bda-d6ec33aaa97c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.984045] env[65385]: DEBUG oslo_vmware.api [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 960.984045] env[65385]: value = "task-4454077" [ 960.984045] env[65385]: _type = "Task" [ 960.984045] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.996886] env[65385]: DEBUG oslo_vmware.api [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.006848] env[65385]: DEBUG nova.network.neutron [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Updating instance_info_cache with network_info: [{"id": "8d177373-bd8d-475c-81b6-972d8eae4f95", "address": "fa:16:3e:2c:a6:49", "network": {"id": "275dc47c-6cdc-4e1d-8629-31f4d73418dd", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1045504663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ab35da093ff491fa13580ad2b1e610f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d177373-bd", "ovs_interfaceid": "8d177373-bd8d-475c-81b6-972d8eae4f95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 961.373848] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.985s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.376203] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454074, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075398} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.376774] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.796s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.378374] env[65385]: INFO nova.compute.claims [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.381149] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.382306] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31516fe-f2cd-43f2-9e5f-71ec3105855f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.410165] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] f4af288e-b9c9-4273-8e65-e726ae039d03/f4af288e-b9c9-4273-8e65-e726ae039d03.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.410792] env[65385]: INFO nova.scheduler.client.report [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted allocations for instance c18b5bde-1a17-4954-a2c5-582815ebfec8 [ 961.415560] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2aad6da-a6dd-49ac-a8ed-158f58ec0512 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.435130] env[65385]: DEBUG nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 961.444440] env[65385]: DEBUG oslo_vmware.api [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183437} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.446219] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.446407] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.448041] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.448041] env[65385]: INFO nova.compute.manager [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Took 1.22 seconds to destroy the instance on the hypervisor. [ 961.448041] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 961.448041] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 961.448041] env[65385]: value = "task-4454078" [ 961.448041] env[65385]: _type = "Task" [ 961.448041] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.448041] env[65385]: DEBUG nova.compute.manager [-] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 961.448041] env[65385]: DEBUG nova.network.neutron [-] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 961.448041] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.448463] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.448704] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.470197] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 961.470197] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 961.470338] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 961.470508] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 961.470669] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 961.470890] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 961.471167] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.471579] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 961.471628] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 961.471794] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 961.472022] env[65385]: DEBUG nova.virt.hardware [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 961.473152] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bcd7bf5-8b95-47ac-8f52-b42ba57f46ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.483940] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca14bf04-f954-4f0f-8d2a-9b5474bc46f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.506036] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.513567] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Releasing lock "refresh_cache-55936da6-fcdf-4291-9b3a-23fa8b7a56ef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 961.514323] env[65385]: DEBUG nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Instance network_info: |[{"id": "8d177373-bd8d-475c-81b6-972d8eae4f95", "address": "fa:16:3e:2c:a6:49", "network": {"id": "275dc47c-6cdc-4e1d-8629-31f4d73418dd", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1045504663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ab35da093ff491fa13580ad2b1e610f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d177373-bd", "ovs_interfaceid": "8d177373-bd8d-475c-81b6-972d8eae4f95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 961.515201] env[65385]: DEBUG oslo_vmware.api [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14382} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.515731] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:a6:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8d177373-bd8d-475c-81b6-972d8eae4f95', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.524153] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Creating folder: Project (5ab35da093ff491fa13580ad2b1e610f). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 961.524621] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.524886] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.525214] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.525522] env[65385]: INFO nova.compute.manager [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Took 1.21 seconds to destroy the instance on the hypervisor. [ 961.525778] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 961.526407] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-118503e9-9aa0-47c7-9406-b60326702ec5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.528367] env[65385]: DEBUG nova.compute.manager [-] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 961.528476] env[65385]: DEBUG nova.network.neutron [-] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 961.528729] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.529460] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.530383] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.547509] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Created folder: Project (5ab35da093ff491fa13580ad2b1e610f) in parent group-v870881. [ 961.547879] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Creating folder: Instances. Parent ref: group-v871119. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 961.548046] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1186a1b4-4de3-455d-94d9-b3f6dd02b9cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.553187] env[65385]: DEBUG nova.compute.manager [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Received event network-changed-8d177373-bd8d-475c-81b6-972d8eae4f95 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 961.553361] env[65385]: DEBUG nova.compute.manager [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Refreshing instance network info cache due to event network-changed-8d177373-bd8d-475c-81b6-972d8eae4f95. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 961.553624] env[65385]: DEBUG oslo_concurrency.lockutils [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Acquiring lock "refresh_cache-55936da6-fcdf-4291-9b3a-23fa8b7a56ef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.553900] env[65385]: DEBUG oslo_concurrency.lockutils [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Acquired lock "refresh_cache-55936da6-fcdf-4291-9b3a-23fa8b7a56ef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.554149] env[65385]: DEBUG nova.network.neutron [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Refreshing network info cache for port 8d177373-bd8d-475c-81b6-972d8eae4f95 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 961.565480] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Created folder: Instances in parent group-v871119. [ 961.565780] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 961.566271] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.566562] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d6a6c81-6e1c-4f99-9187-9c361aa79d7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.588912] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.588912] env[65385]: value = "task-4454081" [ 961.588912] env[65385]: _type = "Task" [ 961.588912] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.599392] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454081, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.811104] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.886841] env[65385]: DEBUG nova.compute.manager [req-e7082c95-7018-4d34-a09b-b8728d005db3 req-06ba1786-69dc-43cc-87d8-3140b3455879 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Received event network-vif-deleted-978b2387-8e83-4565-9a81-983a73480cf1 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 961.886841] env[65385]: INFO nova.compute.manager [req-e7082c95-7018-4d34-a09b-b8728d005db3 req-06ba1786-69dc-43cc-87d8-3140b3455879 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Neutron deleted interface 978b2387-8e83-4565-9a81-983a73480cf1; detaching it from the instance and deleting it from the info cache [ 961.887044] env[65385]: DEBUG nova.network.neutron [req-e7082c95-7018-4d34-a09b-b8728d005db3 req-06ba1786-69dc-43cc-87d8-3140b3455879 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 961.940271] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85f548de-9bc3-46c2-a1e3-933e0ebdd304 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "c18b5bde-1a17-4954-a2c5-582815ebfec8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.967s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.967906] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454078, 'name': ReconfigVM_Task, 'duration_secs': 0.338206} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.968238] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Reconfigured VM instance instance-00000053 to attach disk [datastore1] f4af288e-b9c9-4273-8e65-e726ae039d03/f4af288e-b9c9-4273-8e65-e726ae039d03.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.969019] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-005bef6d-1ff8-434c-9104-7afbbd9b9f70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.983527] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 961.983527] env[65385]: value = "task-4454082" [ 961.983527] env[65385]: _type = "Task" [ 961.983527] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.989392] env[65385]: DEBUG nova.network.neutron [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Successfully updated port: 82a01c37-b04c-4b2c-a613-dc4117b64ef2 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 961.992475] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454082, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.057926] env[65385]: WARNING neutronclient.v2_0.client [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 962.058964] env[65385]: WARNING openstack [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 962.059339] env[65385]: WARNING openstack [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 962.099896] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454081, 'name': CreateVM_Task, 'duration_secs': 0.456171} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.100095] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.100594] env[65385]: WARNING neutronclient.v2_0.client [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 962.100951] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.101212] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.101548] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 962.102179] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd313afe-d0a2-48d4-9309-d677dff5abb0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.108176] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 962.108176] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e692be-e866-bac5-b3e4-ac77f199a130" [ 962.108176] env[65385]: _type = "Task" [ 962.108176] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.117464] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e692be-e866-bac5-b3e4-ac77f199a130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.175687] env[65385]: WARNING openstack [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 962.176411] env[65385]: WARNING openstack [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 962.249672] env[65385]: WARNING neutronclient.v2_0.client [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 962.250482] env[65385]: WARNING openstack [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 962.250903] env[65385]: WARNING openstack [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 962.280467] env[65385]: DEBUG nova.network.neutron [-] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 962.340299] env[65385]: DEBUG nova.network.neutron [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Updated VIF entry in instance network info cache for port 8d177373-bd8d-475c-81b6-972d8eae4f95. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 962.340671] env[65385]: DEBUG nova.network.neutron [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Updating instance_info_cache with network_info: [{"id": "8d177373-bd8d-475c-81b6-972d8eae4f95", "address": "fa:16:3e:2c:a6:49", "network": {"id": "275dc47c-6cdc-4e1d-8629-31f4d73418dd", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1045504663-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ab35da093ff491fa13580ad2b1e610f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d177373-bd", "ovs_interfaceid": "8d177373-bd8d-475c-81b6-972d8eae4f95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 962.396267] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b3a0f55-bd62-4a48-89cd-b84ab9c7783e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.406578] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278943c5-e971-4f96-8eb7-bdb6408850c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.443315] env[65385]: DEBUG nova.compute.manager [req-e7082c95-7018-4d34-a09b-b8728d005db3 req-06ba1786-69dc-43cc-87d8-3140b3455879 service nova] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Detach interface failed, port_id=978b2387-8e83-4565-9a81-983a73480cf1, reason: Instance f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 962.495549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-c003dd80-d419-4a5a-8f86-dcac1a1e789c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.495549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-c003dd80-d419-4a5a-8f86-dcac1a1e789c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.495549] env[65385]: DEBUG nova.network.neutron [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 962.496567] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454082, 'name': Rename_Task, 'duration_secs': 0.177336} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.497178] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.497178] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-742f29ba-5029-4d09-a741-b2ca61af2958 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.508539] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 962.508539] env[65385]: value = "task-4454083" [ 962.508539] env[65385]: _type = "Task" [ 962.508539] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.519941] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.577730] env[65385]: DEBUG nova.network.neutron [-] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 962.620523] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e692be-e866-bac5-b3e4-ac77f199a130, 'name': SearchDatastore_Task, 'duration_secs': 0.014414} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.624141] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.624438] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.624727] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.624919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.625154] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.625914] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7f6348b-3cdc-4ec9-8000-cc331bdb2e2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.640126] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.640126] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.640997] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92417f42-d331-4564-ad42-9fc4b1fa4b1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.647684] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 962.647684] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5286546f-8511-8d60-e35c-a5c723f453d4" [ 962.647684] env[65385]: _type = "Task" [ 962.647684] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.657213] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5286546f-8511-8d60-e35c-a5c723f453d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.704142] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f61ddb-3f8f-4ee7-bf7a-706c843628ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.713136] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bde5c8-53c5-4f11-a164-7ad67bde7151 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.747540] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870c5811-f751-4e0d-b323-8ccdb5eab175 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.757035] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf9b101-a325-46c7-8789-ec93492037e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.777259] env[65385]: DEBUG nova.compute.provider_tree [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.782948] env[65385]: INFO nova.compute.manager [-] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Took 1.34 seconds to deallocate network for instance. [ 962.845401] env[65385]: DEBUG oslo_concurrency.lockutils [req-51559ba9-f98a-4830-98cf-0aa3aa501cbe req-13514902-eb9d-4fd9-8acb-9cf34f74f160 service nova] Releasing lock "refresh_cache-55936da6-fcdf-4291-9b3a-23fa8b7a56ef" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.001060] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.001612] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.018950] env[65385]: DEBUG oslo_vmware.api [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454083, 'name': PowerOnVM_Task, 'duration_secs': 0.509112} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.019254] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.019447] env[65385]: INFO nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Took 8.51 seconds to spawn the instance on the hypervisor. [ 963.019644] env[65385]: DEBUG nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 963.020560] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb19376-cfbc-42ec-b9fa-7017a1fb6946 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.040548] env[65385]: DEBUG nova.network.neutron [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 963.065267] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.065505] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.081272] env[65385]: INFO nova.compute.manager [-] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Took 1.55 seconds to deallocate network for instance. [ 963.161432] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5286546f-8511-8d60-e35c-a5c723f453d4, 'name': SearchDatastore_Task, 'duration_secs': 0.011989} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.162603] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6d277bf-1794-4aa4-96da-c2ae0500e3f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.170319] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 963.170319] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290bb8a-be16-e50e-e769-b506f629ed9a" [ 963.170319] env[65385]: _type = "Task" [ 963.170319] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.181015] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290bb8a-be16-e50e-e769-b506f629ed9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.280722] env[65385]: DEBUG nova.scheduler.client.report [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 963.295763] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.329482] env[65385]: WARNING neutronclient.v2_0.client [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.330350] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.330753] env[65385]: WARNING openstack [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.490724] env[65385]: DEBUG nova.network.neutron [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Updating instance_info_cache with network_info: [{"id": "82a01c37-b04c-4b2c-a613-dc4117b64ef2", "address": "fa:16:3e:05:c5:5a", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a01c37-b0", "ovs_interfaceid": "82a01c37-b04c-4b2c-a613-dc4117b64ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 963.541425] env[65385]: INFO nova.compute.manager [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Took 22.67 seconds to build instance. [ 963.590614] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.682724] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290bb8a-be16-e50e-e769-b506f629ed9a, 'name': SearchDatastore_Task, 'duration_secs': 0.019156} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.682988] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.683267] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 55936da6-fcdf-4291-9b3a-23fa8b7a56ef/55936da6-fcdf-4291-9b3a-23fa8b7a56ef.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.683532] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-235be5fb-8bdc-414f-8786-cb05159cee06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.691559] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 963.691559] env[65385]: value = "task-4454084" [ 963.691559] env[65385]: _type = "Task" [ 963.691559] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.700627] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454084, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.786955] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 963.787590] env[65385]: DEBUG nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 963.790347] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 11.720s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.790524] env[65385]: DEBUG nova.objects.instance [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 963.994797] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-c003dd80-d419-4a5a-8f86-dcac1a1e789c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.995235] env[65385]: DEBUG nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Instance network_info: |[{"id": "82a01c37-b04c-4b2c-a613-dc4117b64ef2", "address": "fa:16:3e:05:c5:5a", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a01c37-b0", "ovs_interfaceid": "82a01c37-b04c-4b2c-a613-dc4117b64ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 963.995866] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:c5:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82a01c37-b04c-4b2c-a613-dc4117b64ef2', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.004362] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 964.004720] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 964.004999] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f4398d8-f883-4019-b971-4b8824991297 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.035211] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 964.035211] env[65385]: value = "task-4454085" [ 964.035211] env[65385]: _type = "Task" [ 964.035211] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.045828] env[65385]: DEBUG oslo_concurrency.lockutils [None req-36e032e8-2d1c-4901-a65c-baebd5676cf4 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.189s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.046295] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454085, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.151150] env[65385]: DEBUG nova.compute.manager [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Received event network-vif-plugged-82a01c37-b04c-4b2c-a613-dc4117b64ef2 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 964.151462] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Acquiring lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.152695] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.152695] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.152695] env[65385]: DEBUG nova.compute.manager [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] No waiting events found dispatching network-vif-plugged-82a01c37-b04c-4b2c-a613-dc4117b64ef2 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 964.152695] env[65385]: WARNING nova.compute.manager [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Received unexpected event network-vif-plugged-82a01c37-b04c-4b2c-a613-dc4117b64ef2 for instance with vm_state building and task_state spawning. [ 964.153172] env[65385]: DEBUG nova.compute.manager [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Received event network-changed-82a01c37-b04c-4b2c-a613-dc4117b64ef2 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 964.153172] env[65385]: DEBUG nova.compute.manager [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Refreshing instance network info cache due to event network-changed-82a01c37-b04c-4b2c-a613-dc4117b64ef2. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 964.153172] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Acquiring lock "refresh_cache-c003dd80-d419-4a5a-8f86-dcac1a1e789c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.153347] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Acquired lock "refresh_cache-c003dd80-d419-4a5a-8f86-dcac1a1e789c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.153347] env[65385]: DEBUG nova.network.neutron [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Refreshing network info cache for port 82a01c37-b04c-4b2c-a613-dc4117b64ef2 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 964.204934] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454084, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.237059] env[65385]: DEBUG nova.compute.manager [req-79df39f2-0f25-4286-827d-883aab0f4f90 req-ff101a2d-0cbe-41cc-81de-c3ae5c52ee65 service nova] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Received event network-vif-deleted-67be7e47-2dcd-4602-8638-fccd0da10003 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 964.295494] env[65385]: DEBUG nova.compute.utils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 964.300319] env[65385]: DEBUG nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 964.300584] env[65385]: DEBUG nova.network.neutron [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 964.301563] env[65385]: WARNING neutronclient.v2_0.client [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 964.301563] env[65385]: WARNING neutronclient.v2_0.client [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 964.302496] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.302950] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.312879] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1e39c054-204b-4d46-8395-71de62e510f3 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.522s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.314734] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.639s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.314994] env[65385]: DEBUG nova.objects.instance [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lazy-loading 'resources' on Instance uuid 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.361321] env[65385]: DEBUG nova.policy [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb4e56b8e92a4167b887a1b8e97e2ebe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f433a71c93fe461a9c27bb729956a8f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 964.550320] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454085, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.657836] env[65385]: WARNING neutronclient.v2_0.client [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 964.658653] env[65385]: WARNING openstack [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.659082] env[65385]: WARNING openstack [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.704129] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454084, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.740666} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.704540] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 55936da6-fcdf-4291-9b3a-23fa8b7a56ef/55936da6-fcdf-4291-9b3a-23fa8b7a56ef.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 964.704757] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.705254] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d6ec263-6ce4-4eef-95e8-1e8b64414298 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.714937] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 964.714937] env[65385]: value = "task-4454086" [ 964.714937] env[65385]: _type = "Task" [ 964.714937] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.728536] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454086, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.801612] env[65385]: DEBUG nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 964.808179] env[65385]: DEBUG nova.network.neutron [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Successfully created port: 06080c91-c437-4033-b17e-28344fc48b5a {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 964.858330] env[65385]: WARNING openstack [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.859125] env[65385]: WARNING openstack [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.955600] env[65385]: WARNING neutronclient.v2_0.client [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 964.956642] env[65385]: WARNING openstack [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.957935] env[65385]: WARNING openstack [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.058315] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454085, 'name': CreateVM_Task, 'duration_secs': 0.631294} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.058555] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 965.059114] env[65385]: WARNING neutronclient.v2_0.client [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.059491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.059642] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 965.059988] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 965.060255] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ef70ebe-c261-41cf-888b-c96479201b53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.066414] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 965.066414] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a43811-8870-8f8e-1a53-220039993edc" [ 965.066414] env[65385]: _type = "Task" [ 965.066414] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.084206] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a43811-8870-8f8e-1a53-220039993edc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.134063] env[65385]: DEBUG nova.compute.manager [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 965.137306] env[65385]: DEBUG nova.network.neutron [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Updated VIF entry in instance network info cache for port 82a01c37-b04c-4b2c-a613-dc4117b64ef2. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 965.137796] env[65385]: DEBUG nova.network.neutron [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Updating instance_info_cache with network_info: [{"id": "82a01c37-b04c-4b2c-a613-dc4117b64ef2", "address": "fa:16:3e:05:c5:5a", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82a01c37-b0", "ovs_interfaceid": "82a01c37-b04c-4b2c-a613-dc4117b64ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 965.140893] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebface40-23f2-407d-83f4-cf0d5c22dedf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.173945] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8080e653-013f-4508-be2f-096581e1bfab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.183415] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9acdf11-2fbb-4fe0-b853-6ff1ac950ac1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.221678] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e448e95d-d56f-4e1f-9045-2233b755ce1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.227668] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.227887] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.238913] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454086, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120443} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.240792] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588923e5-e55b-4fdb-8922-352b19a701b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.245041] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.246345] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a343251d-c35f-43d8-b547-2151a01bb302 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.262382] env[65385]: DEBUG nova.compute.provider_tree [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.282213] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 55936da6-fcdf-4291-9b3a-23fa8b7a56ef/55936da6-fcdf-4291-9b3a-23fa8b7a56ef.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.283387] env[65385]: DEBUG nova.scheduler.client.report [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 965.286779] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69b4248e-4566-45ef-9ef3-d7ef2e74747b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.304401] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.990s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.306463] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.380s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.307961] env[65385]: INFO nova.compute.claims [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 965.318237] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 965.318237] env[65385]: value = "task-4454087" [ 965.318237] env[65385]: _type = "Task" [ 965.318237] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.332435] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454087, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.335256] env[65385]: INFO nova.scheduler.client.report [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Deleted allocations for instance 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b [ 965.579234] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a43811-8870-8f8e-1a53-220039993edc, 'name': SearchDatastore_Task, 'duration_secs': 0.017349} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.579608] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 965.579837] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 965.580139] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.580334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 965.580524] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.580850] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-635a5058-625f-48e6-af36-c8142b2e35bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.594074] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.594074] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 965.594309] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80050afa-c449-4eda-8760-5375b0512885 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.600903] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 965.600903] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527562fe-916a-6d01-228f-b22723a5176d" [ 965.600903] env[65385]: _type = "Task" [ 965.600903] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.609970] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527562fe-916a-6d01-228f-b22723a5176d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.645770] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5f933af-1d70-441e-9b19-51a9ce9947c7 req-4acfd33f-6503-407a-a456-1f73da2f44a6 service nova] Releasing lock "refresh_cache-c003dd80-d419-4a5a-8f86-dcac1a1e789c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 965.654203] env[65385]: INFO nova.compute.manager [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] instance snapshotting [ 965.657714] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee9ec54-0bd6-49bc-9cce-771322ef2cc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.680362] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c8a1a5-41f8-4052-b1eb-3bfb6adf4482 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.731182] env[65385]: DEBUG nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 965.817827] env[65385]: DEBUG nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 965.835056] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454087, 'name': ReconfigVM_Task, 'duration_secs': 0.400506} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.835056] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 55936da6-fcdf-4291-9b3a-23fa8b7a56ef/55936da6-fcdf-4291-9b3a-23fa8b7a56ef.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.835778] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b7ea81e-4b65-48c6-bf9e-0b0d3421b348 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.844394] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 965.844394] env[65385]: value = "task-4454088" [ 965.844394] env[65385]: _type = "Task" [ 965.844394] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.844679] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3536a267-9904-465c-a0ab-2f50100b2c2f tempest-ServerMetadataTestJSON-770388445 tempest-ServerMetadataTestJSON-770388445-project-member] Lock "78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.658s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.847753] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 965.847971] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 965.848136] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 965.848324] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 965.848464] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 965.848605] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 965.848807] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.848961] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 965.849135] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 965.849298] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 965.849465] env[65385]: DEBUG nova.virt.hardware [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 965.857025] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc6c70d-c15a-418e-a4b2-9bdb0550ea40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.868378] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454088, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.871948] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe93e83-ca7a-48e7-9abc-f44833a4eb98 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.112413] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527562fe-916a-6d01-228f-b22723a5176d, 'name': SearchDatastore_Task, 'duration_secs': 0.01629} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.113292] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b8a79d9-71f4-4e4c-b0f4-3c4dff0c30a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.119505] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 966.119505] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520473f1-4716-68fa-6feb-98f81f7b24b8" [ 966.119505] env[65385]: _type = "Task" [ 966.119505] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.128706] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520473f1-4716-68fa-6feb-98f81f7b24b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.194320] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 966.196249] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f484a2ff-a0aa-4f93-9e86-7ca023477ac9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.204220] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 966.204220] env[65385]: value = "task-4454089" [ 966.204220] env[65385]: _type = "Task" [ 966.204220] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.216181] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454089, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.260314] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.357586] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454088, 'name': Rename_Task, 'duration_secs': 0.199751} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.357923] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.358216] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40810d72-9bdb-43c5-b001-a95fe61a6a5e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.368714] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 966.368714] env[65385]: value = "task-4454090" [ 966.368714] env[65385]: _type = "Task" [ 966.368714] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.379039] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.380250] env[65385]: DEBUG nova.network.neutron [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Successfully updated port: 06080c91-c437-4033-b17e-28344fc48b5a {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 966.607052] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cde0b9-9d3c-446c-aaf8-597f6321aaab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.617633] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a25a9e-3e36-4772-8f6e-de493010f6bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.622916] env[65385]: DEBUG nova.compute.manager [req-0149a929-c6c5-45c3-8929-39263f8919ee req-739d9d27-48ab-4cf1-aed3-e2e730e259b5 service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Received event network-vif-plugged-06080c91-c437-4033-b17e-28344fc48b5a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 966.623190] env[65385]: DEBUG oslo_concurrency.lockutils [req-0149a929-c6c5-45c3-8929-39263f8919ee req-739d9d27-48ab-4cf1-aed3-e2e730e259b5 service nova] Acquiring lock "167e1b46-af52-4a7a-9964-edabc8155287-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.623413] env[65385]: DEBUG oslo_concurrency.lockutils [req-0149a929-c6c5-45c3-8929-39263f8919ee req-739d9d27-48ab-4cf1-aed3-e2e730e259b5 service nova] Lock "167e1b46-af52-4a7a-9964-edabc8155287-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.623620] env[65385]: DEBUG oslo_concurrency.lockutils [req-0149a929-c6c5-45c3-8929-39263f8919ee req-739d9d27-48ab-4cf1-aed3-e2e730e259b5 service nova] Lock "167e1b46-af52-4a7a-9964-edabc8155287-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.623817] env[65385]: DEBUG nova.compute.manager [req-0149a929-c6c5-45c3-8929-39263f8919ee req-739d9d27-48ab-4cf1-aed3-e2e730e259b5 service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] No waiting events found dispatching network-vif-plugged-06080c91-c437-4033-b17e-28344fc48b5a {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 966.623992] env[65385]: WARNING nova.compute.manager [req-0149a929-c6c5-45c3-8929-39263f8919ee req-739d9d27-48ab-4cf1-aed3-e2e730e259b5 service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Received unexpected event network-vif-plugged-06080c91-c437-4033-b17e-28344fc48b5a for instance with vm_state building and task_state spawning. [ 966.669727] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520473f1-4716-68fa-6feb-98f81f7b24b8, 'name': SearchDatastore_Task, 'duration_secs': 0.019304} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.670537] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f12415-68f1-44b7-bdf3-c53c543a7722 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.673711] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 966.674018] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c003dd80-d419-4a5a-8f86-dcac1a1e789c/c003dd80-d419-4a5a-8f86-dcac1a1e789c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 966.674554] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d0360fb-43a3-4aa1-8906-e7ecc0aad229 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.685340] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a07baa8-7e25-405d-908a-038cf890beba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.689144] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 966.689144] env[65385]: value = "task-4454091" [ 966.689144] env[65385]: _type = "Task" [ 966.689144] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.705403] env[65385]: DEBUG nova.compute.provider_tree [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.722985] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454091, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.729926] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454089, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.882254] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454090, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.887966] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "refresh_cache-167e1b46-af52-4a7a-9964-edabc8155287" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.888199] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquired lock "refresh_cache-167e1b46-af52-4a7a-9964-edabc8155287" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 966.888415] env[65385]: DEBUG nova.network.neutron [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 967.203413] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454091, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.215788] env[65385]: DEBUG nova.scheduler.client.report [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 967.227912] env[65385]: INFO nova.compute.manager [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Rebuilding instance [ 967.230510] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454089, 'name': CreateSnapshot_Task, 'duration_secs': 0.824646} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.232074] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 967.236522] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf62d94f-86a0-4e88-aaf0-011da9c6419c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.294925] env[65385]: DEBUG nova.compute.manager [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 967.296880] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0f55f0-5ba9-4dea-83ea-c5f135d05f78 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.384163] env[65385]: DEBUG oslo_vmware.api [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454090, 'name': PowerOnVM_Task, 'duration_secs': 0.702454} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.384424] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.384638] env[65385]: INFO nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Took 8.40 seconds to spawn the instance on the hypervisor. [ 967.384832] env[65385]: DEBUG nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 967.385785] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ed2d2f-24f3-4eeb-bff2-4a9806824393 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.392350] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.393070] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.432565] env[65385]: DEBUG nova.network.neutron [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 967.452377] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.452782] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.470940] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ed37a-7507-25de-b4f6-d6b081440980/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 967.471939] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8ab964-d80f-44e7-9c47-b1fcedc7c9cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.479274] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ed37a-7507-25de-b4f6-d6b081440980/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 967.479462] env[65385]: ERROR oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ed37a-7507-25de-b4f6-d6b081440980/disk-0.vmdk due to incomplete transfer. [ 967.479704] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e27bf9a3-11fe-46e1-9113-163f7067a882 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.490935] env[65385]: DEBUG oslo_vmware.rw_handles [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527ed37a-7507-25de-b4f6-d6b081440980/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 967.491051] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Uploaded image 1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 967.495379] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 967.495379] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-17ed0706-9f03-4360-aed6-8bbfaf94819a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.501907] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 967.501907] env[65385]: value = "task-4454092" [ 967.501907] env[65385]: _type = "Task" [ 967.501907] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.511751] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454092, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.542586] env[65385]: WARNING neutronclient.v2_0.client [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.543045] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.543462] env[65385]: WARNING openstack [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.693798] env[65385]: DEBUG nova.network.neutron [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Updating instance_info_cache with network_info: [{"id": "06080c91-c437-4033-b17e-28344fc48b5a", "address": "fa:16:3e:7a:70:a5", "network": {"id": "2d55ec7e-01d5-4197-a6e5-e70f2099cabd", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1403394817-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f433a71c93fe461a9c27bb729956a8f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06080c91-c4", "ovs_interfaceid": "06080c91-c437-4033-b17e-28344fc48b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 967.711977] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454091, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.665169} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.711977] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c003dd80-d419-4a5a-8f86-dcac1a1e789c/c003dd80-d419-4a5a-8f86-dcac1a1e789c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 967.712447] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 967.712981] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6df92063-4601-4bed-8923-b5950ca24d71 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.725769] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 967.725769] env[65385]: value = "task-4454093" [ 967.725769] env[65385]: _type = "Task" [ 967.725769] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.733105] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.734149] env[65385]: DEBUG nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 967.744650] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.619s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.745291] env[65385]: DEBUG nova.objects.instance [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lazy-loading 'resources' on Instance uuid 0c345794-036c-45d1-985e-cf5c116bd873 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.747114] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454093, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.765876] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 967.767558] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d47654d4-f044-4052-a95c-ba5a80f63491 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.780851] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 967.780851] env[65385]: value = "task-4454094" [ 967.780851] env[65385]: _type = "Task" [ 967.780851] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.794798] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454094, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.917116] env[65385]: INFO nova.compute.manager [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Took 23.19 seconds to build instance. [ 968.017034] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454092, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.202308] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Releasing lock "refresh_cache-167e1b46-af52-4a7a-9964-edabc8155287" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 968.202698] env[65385]: DEBUG nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Instance network_info: |[{"id": "06080c91-c437-4033-b17e-28344fc48b5a", "address": "fa:16:3e:7a:70:a5", "network": {"id": "2d55ec7e-01d5-4197-a6e5-e70f2099cabd", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1403394817-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f433a71c93fe461a9c27bb729956a8f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06080c91-c4", "ovs_interfaceid": "06080c91-c437-4033-b17e-28344fc48b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 968.203301] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:70:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06080c91-c437-4033-b17e-28344fc48b5a', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.211061] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Creating folder: Project (f433a71c93fe461a9c27bb729956a8f2). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 968.211446] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d6777e1-d45d-412b-991b-8b591c6e8424 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.224804] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Created folder: Project (f433a71c93fe461a9c27bb729956a8f2) in parent group-v870881. [ 968.225037] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Creating folder: Instances. Parent ref: group-v871125. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 968.225307] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6e6a935-0794-416e-9f09-0f6a430ae57c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.239695] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454093, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.249214} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.239974] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.240779] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b4a0ad-7b71-4438-abcc-cfa21f302b3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.244792] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Created folder: Instances in parent group-v871125. [ 968.245017] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 968.245618] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 968.245919] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97083a45-5a88-46b0-a6b6-387435f4606f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.273711] env[65385]: DEBUG nova.compute.utils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 968.275471] env[65385]: DEBUG nova.objects.instance [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lazy-loading 'numa_topology' on Instance uuid 0c345794-036c-45d1-985e-cf5c116bd873 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.285608] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] c003dd80-d419-4a5a-8f86-dcac1a1e789c/c003dd80-d419-4a5a-8f86-dcac1a1e789c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.286627] env[65385]: DEBUG nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 968.287616] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68b53af6-5613-4430-ae72-f0c646840633 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.309926] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.309926] env[65385]: value = "task-4454097" [ 968.309926] env[65385]: _type = "Task" [ 968.309926] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.323084] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.323084] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454094, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.323084] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b479f5ad-f10b-44db-8f17-bead4e0ee42a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.325488] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 968.325488] env[65385]: value = "task-4454098" [ 968.325488] env[65385]: _type = "Task" [ 968.325488] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.335333] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454097, 'name': CreateVM_Task} progress is 15%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.341158] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454098, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.341651] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 968.341651] env[65385]: value = "task-4454099" [ 968.341651] env[65385]: _type = "Task" [ 968.341651] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.354813] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.411303] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.420331] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c7b34fc4-71ab-4066-8dc2-2feeafbdd2d3 tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.701s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 968.420842] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.010s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 968.421163] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.421400] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 968.421583] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 968.424196] env[65385]: INFO nova.compute.manager [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Terminating instance [ 968.522023] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454092, 'name': Destroy_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.648640] env[65385]: DEBUG nova.compute.manager [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Received event network-changed-06080c91-c437-4033-b17e-28344fc48b5a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 968.648793] env[65385]: DEBUG nova.compute.manager [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Refreshing instance network info cache due to event network-changed-06080c91-c437-4033-b17e-28344fc48b5a. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 968.648974] env[65385]: DEBUG oslo_concurrency.lockutils [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Acquiring lock "refresh_cache-167e1b46-af52-4a7a-9964-edabc8155287" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.649133] env[65385]: DEBUG oslo_concurrency.lockutils [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Acquired lock "refresh_cache-167e1b46-af52-4a7a-9964-edabc8155287" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 968.649290] env[65385]: DEBUG nova.network.neutron [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Refreshing network info cache for port 06080c91-c437-4033-b17e-28344fc48b5a {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 968.787613] env[65385]: DEBUG nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 968.790873] env[65385]: DEBUG nova.objects.base [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Object Instance<0c345794-036c-45d1-985e-cf5c116bd873> lazy-loaded attributes: resources,numa_topology {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 968.804758] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454094, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.822017] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454097, 'name': CreateVM_Task, 'duration_secs': 0.384182} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.822150] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.822998] env[65385]: WARNING neutronclient.v2_0.client [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 968.823120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.823287] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 968.823730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 968.825030] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1e4e6bc-2c96-4c0c-8c36-26da023b6df3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.832145] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 968.832145] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aad93d-844d-997f-f4c9-740d728e19c0" [ 968.832145] env[65385]: _type = "Task" [ 968.832145] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.844966] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.853649] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aad93d-844d-997f-f4c9-740d728e19c0, 'name': SearchDatastore_Task, 'duration_secs': 0.01243} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.854426] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 968.854662] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.854950] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.855183] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 968.855447] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.858755] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b835443-f375-43b7-b953-6019cf86fc06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.865051] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454099, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.875338] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.875629] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.876364] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebe5d190-8333-4fa6-90c9-3a855b12186f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.882844] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 968.882844] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eb3dd8-20a6-63db-da42-012097282a4f" [ 968.882844] env[65385]: _type = "Task" [ 968.882844] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.894587] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eb3dd8-20a6-63db-da42-012097282a4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.928646] env[65385]: DEBUG nova.compute.manager [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 968.928873] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.929734] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165dfe2c-1aa7-499d-a6f1-0cd7dbca2f42 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.938753] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.941597] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbac9cf8-a08c-494a-93d4-16690c63cc92 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.951039] env[65385]: DEBUG oslo_vmware.api [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 968.951039] env[65385]: value = "task-4454100" [ 968.951039] env[65385]: _type = "Task" [ 968.951039] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.962882] env[65385]: DEBUG oslo_vmware.api [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454100, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.016232] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454092, 'name': Destroy_Task, 'duration_secs': 1.017739} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.016591] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Destroyed the VM [ 969.016949] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 969.017333] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b282dc7c-1028-4e0f-abb8-4a27bdde5bed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.030444] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 969.030444] env[65385]: value = "task-4454101" [ 969.030444] env[65385]: _type = "Task" [ 969.030444] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.041265] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454101, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.082016] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c79fee-a286-4009-af9c-aeaac59993cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.090735] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5aed02-0e18-47a5-83cd-e14190c80dd3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.128138] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497dbaba-bf36-4927-b0b4-94331953a215 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.138476] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d82bdd-0529-42ad-9985-103d0db74ea8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.155800] env[65385]: WARNING neutronclient.v2_0.client [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.156652] env[65385]: WARNING openstack [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.156964] env[65385]: WARNING openstack [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.165350] env[65385]: DEBUG nova.compute.provider_tree [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.270301] env[65385]: WARNING openstack [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.270778] env[65385]: WARNING openstack [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.320592] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454094, 'name': CloneVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.338040] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454098, 'name': ReconfigVM_Task, 'duration_secs': 0.5854} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.338040] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Reconfigured VM instance instance-00000055 to attach disk [datastore2] c003dd80-d419-4a5a-8f86-dcac1a1e789c/c003dd80-d419-4a5a-8f86-dcac1a1e789c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.338666] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-edce6998-c191-443b-a6d1-07a65454cfc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.350651] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 969.350651] env[65385]: value = "task-4454102" [ 969.350651] env[65385]: _type = "Task" [ 969.350651] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.351944] env[65385]: WARNING neutronclient.v2_0.client [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.352693] env[65385]: WARNING openstack [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.353057] env[65385]: WARNING openstack [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.373695] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454099, 'name': PowerOffVM_Task, 'duration_secs': 0.555648} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.374460] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.374792] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.375695] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c7cc1d-09cd-4d8c-b849-fd996399c1b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.384544] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454102, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.392717] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.393470] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4383d6f-1af6-4510-83ec-bf8552ee76ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.400513] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eb3dd8-20a6-63db-da42-012097282a4f, 'name': SearchDatastore_Task, 'duration_secs': 0.011369} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.401446] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5a90818-7f0f-41dc-8eb9-855b477a6fb3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.409216] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 969.409216] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284b689-ef41-868e-db19-d3da14055cf3" [ 969.409216] env[65385]: _type = "Task" [ 969.409216] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.420850] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284b689-ef41-868e-db19-d3da14055cf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.462239] env[65385]: DEBUG oslo_vmware.api [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454100, 'name': PowerOffVM_Task, 'duration_secs': 0.220082} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.462597] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.462782] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.463139] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6b4121a-3b65-4b7a-8a87-8f30cb7e453c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.468996] env[65385]: DEBUG nova.network.neutron [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Updated VIF entry in instance network info cache for port 06080c91-c437-4033-b17e-28344fc48b5a. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 969.469432] env[65385]: DEBUG nova.network.neutron [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Updating instance_info_cache with network_info: [{"id": "06080c91-c437-4033-b17e-28344fc48b5a", "address": "fa:16:3e:7a:70:a5", "network": {"id": "2d55ec7e-01d5-4197-a6e5-e70f2099cabd", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1403394817-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f433a71c93fe461a9c27bb729956a8f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06080c91-c4", "ovs_interfaceid": "06080c91-c437-4033-b17e-28344fc48b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 969.472323] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.472627] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.472901] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleting the datastore file [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.473516] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e888921-6a7c-420c-a082-c76ff9d4eea2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.482430] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 969.482430] env[65385]: value = "task-4454105" [ 969.482430] env[65385]: _type = "Task" [ 969.482430] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.493991] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454105, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.539305] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.539544] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.539723] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Deleting the datastore file [datastore1] 55936da6-fcdf-4291-9b3a-23fa8b7a56ef {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.543454] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8986c3e-7e96-45a2-9fb8-03a1e7466f0f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.545666] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454101, 'name': RemoveSnapshot_Task, 'duration_secs': 0.412783} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.545927] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 969.546220] env[65385]: DEBUG nova.compute.manager [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 969.547552] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259f2a2e-c7e5-4a45-8551-9e0b80540906 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.551819] env[65385]: DEBUG oslo_vmware.api [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for the task: (returnval){ [ 969.551819] env[65385]: value = "task-4454106" [ 969.551819] env[65385]: _type = "Task" [ 969.551819] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.567135] env[65385]: DEBUG oslo_vmware.api [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.668921] env[65385]: DEBUG nova.scheduler.client.report [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 969.804839] env[65385]: DEBUG nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 969.818077] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454094, 'name': CloneVM_Task, 'duration_secs': 1.632545} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.818375] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Created linked-clone VM from snapshot [ 969.819236] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83423933-35f3-4fe6-8827-659ee54db9e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.828566] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Uploading image 034b156e-8cd8-44b1-b387-b759f79925e6 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 969.839499] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 969.839898] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 969.840194] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 969.840410] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 969.840566] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 969.840736] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 969.840975] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.841169] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 969.841340] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 969.841543] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 969.841721] env[65385]: DEBUG nova.virt.hardware [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 969.842636] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df62f1bb-7593-4637-964a-eb634893fc25 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.846338] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 969.846583] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e58ab96e-78e8-41dc-bf08-e13d8d1ae407 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.854777] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b76aa8-3632-4022-9f2d-2037789c2ba9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.860088] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 969.860088] env[65385]: value = "task-4454107" [ 969.860088] env[65385]: _type = "Task" [ 969.860088] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.873611] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.879092] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Creating folder: Project (cd044bcae7ac489ab0ac7e264c4b5c21). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.882716] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e90908c0-5e3a-4e6c-b776-575f0a00575c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.889284] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454107, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.892655] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454102, 'name': Rename_Task, 'duration_secs': 0.157818} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.892954] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.893247] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-743a2558-2d38-4359-874a-82471647188b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.896292] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Created folder: Project (cd044bcae7ac489ab0ac7e264c4b5c21) in parent group-v870881. [ 969.896395] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Creating folder: Instances. Parent ref: group-v871128. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.897042] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-147c5611-8169-49a6-9063-cb6f30328c86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.902376] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 969.902376] env[65385]: value = "task-4454109" [ 969.902376] env[65385]: _type = "Task" [ 969.902376] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.909267] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Created folder: Instances in parent group-v871128. [ 969.909630] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 969.913534] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.913846] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.918305] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98fdf75a-b0bc-4c65-9580-d18cd8d526b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.937742] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5284b689-ef41-868e-db19-d3da14055cf3, 'name': SearchDatastore_Task, 'duration_secs': 0.012986} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.940733] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 969.940733] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 167e1b46-af52-4a7a-9964-edabc8155287/167e1b46-af52-4a7a-9964-edabc8155287.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.940733] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.940733] env[65385]: value = "task-4454111" [ 969.940733] env[65385]: _type = "Task" [ 969.940733] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.940733] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-792f6259-411c-4917-8d49-936ff98117fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.953845] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454111, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.956149] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 969.956149] env[65385]: value = "task-4454112" [ 969.956149] env[65385]: _type = "Task" [ 969.956149] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.967012] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.974942] env[65385]: DEBUG oslo_concurrency.lockutils [req-c94ea611-a218-45db-aef5-dfeeeae71ee7 req-d1bc3f7b-8f3f-48f9-bfba-8ebb3a1919be service nova] Releasing lock "refresh_cache-167e1b46-af52-4a7a-9964-edabc8155287" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 969.996189] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454105, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.347398} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.996462] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.996645] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.996827] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.070029] env[65385]: DEBUG oslo_vmware.api [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Task: {'id': task-4454106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.340472} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.070870] env[65385]: INFO nova.compute.manager [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Shelve offloading [ 970.071884] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.072080] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.072294] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.072479] env[65385]: INFO nova.compute.manager [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Took 1.14 seconds to destroy the instance on the hypervisor. [ 970.072700] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 970.072902] env[65385]: DEBUG nova.compute.manager [-] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 970.073029] env[65385]: DEBUG nova.network.neutron [-] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 970.073588] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.074281] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.074390] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.116830] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.174495] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.430s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.177700] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.135s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.178213] env[65385]: DEBUG nova.objects.instance [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'resources' on Instance uuid 4b0b8859-b63e-4740-bf67-5733c75e1c70 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.373115] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454107, 'name': Destroy_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.415829] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454109, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.456014] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454111, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.469146] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454112, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50811} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.469584] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 167e1b46-af52-4a7a-9964-edabc8155287/167e1b46-af52-4a7a-9964-edabc8155287.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.469868] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.470226] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6cfe056b-8bab-403d-bbc3-4debef2357d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.480066] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 970.480066] env[65385]: value = "task-4454113" [ 970.480066] env[65385]: _type = "Task" [ 970.480066] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.488827] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.585082] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.585498] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-195e7be8-7dd8-4659-82bc-c065e9345e63 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.593897] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 970.593897] env[65385]: value = "task-4454114" [ 970.593897] env[65385]: _type = "Task" [ 970.593897] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.604736] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.687803] env[65385]: DEBUG nova.compute.manager [req-886206d4-4ea8-4fd6-8d3c-aa0ee89b039f req-9ca1334f-73d1-4a3c-bcb4-7fbd318948a6 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Received event network-vif-deleted-8d177373-bd8d-475c-81b6-972d8eae4f95 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 970.687803] env[65385]: INFO nova.compute.manager [req-886206d4-4ea8-4fd6-8d3c-aa0ee89b039f req-9ca1334f-73d1-4a3c-bcb4-7fbd318948a6 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Neutron deleted interface 8d177373-bd8d-475c-81b6-972d8eae4f95; detaching it from the instance and deleting it from the info cache [ 970.687803] env[65385]: DEBUG nova.network.neutron [req-886206d4-4ea8-4fd6-8d3c-aa0ee89b039f req-9ca1334f-73d1-4a3c-bcb4-7fbd318948a6 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 970.691648] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a8093378-ad53-41ea-8845-697b39c30e2b tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 36.277s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.692744] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 13.432s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.692986] env[65385]: INFO nova.compute.manager [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Unshelving [ 970.871970] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454107, 'name': Destroy_Task, 'duration_secs': 0.775804} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.874881] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Destroyed the VM [ 970.875165] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 970.875617] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4b7aa27d-be3a-4213-96c9-f5dc85c902d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.877364] env[65385]: DEBUG nova.network.neutron [-] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 970.886680] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 970.886680] env[65385]: value = "task-4454115" [ 970.886680] env[65385]: _type = "Task" [ 970.886680] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.899805] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454115, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.914148] env[65385]: DEBUG oslo_vmware.api [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454109, 'name': PowerOnVM_Task, 'duration_secs': 0.804429} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.914447] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.914684] env[65385]: INFO nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Took 9.48 seconds to spawn the instance on the hypervisor. [ 970.914876] env[65385]: DEBUG nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 970.915689] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66e8a5b-a9f4-4190-bea6-47ff666db621 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.945497] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84547c0-efb9-4fb1-9fbe-6c885ec6c2e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.960623] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b9b415-6074-4d1e-9d4d-7ffd14a4b31a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.964172] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454111, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.997258] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdfba5b-be18-4141-a337-3f9a80381a61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.009212] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758ab5f8-a3b5-41cd-a20f-86fc4e5950a0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.016306] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.027994] env[65385]: DEBUG nova.compute.provider_tree [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.047847] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 971.048166] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 971.048459] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 971.048681] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 971.048763] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 971.048851] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 971.049073] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.049260] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 971.049433] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 971.049591] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 971.049775] env[65385]: DEBUG nova.virt.hardware [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 971.052078] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509d54a1-0760-49fb-b0b9-db9f4258cd6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.064874] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69ffd2d-12fa-4ae7-a196-150da5463c39 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.082213] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:5f:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '722cc380-0268-4488-ad6c-462f3af7b938', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.089853] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 971.090216] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.090450] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69a2a913-84e1-4156-b218-5dc852ed5654 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.116713] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 971.116978] env[65385]: DEBUG nova.compute.manager [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 971.117260] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.117260] env[65385]: value = "task-4454116" [ 971.117260] env[65385]: _type = "Task" [ 971.117260] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.118024] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d80464f-02b3-4819-957a-aa7a465bb333 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.133471] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.133762] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.133953] env[65385]: DEBUG nova.network.neutron [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 971.135608] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454116, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.193520] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb69659b-8733-47cb-a13d-f46a836e5284 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.211714] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd197f6-141c-47b1-aac7-0dce3ab1bdd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.248428] env[65385]: DEBUG nova.compute.manager [req-886206d4-4ea8-4fd6-8d3c-aa0ee89b039f req-9ca1334f-73d1-4a3c-bcb4-7fbd318948a6 service nova] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Detach interface failed, port_id=8d177373-bd8d-475c-81b6-972d8eae4f95, reason: Instance 55936da6-fcdf-4291-9b3a-23fa8b7a56ef could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 971.381873] env[65385]: INFO nova.compute.manager [-] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Took 1.31 seconds to deallocate network for instance. [ 971.401620] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454115, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.437168] env[65385]: INFO nova.compute.manager [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Took 24.80 seconds to build instance. [ 971.455502] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454111, 'name': CreateVM_Task, 'duration_secs': 1.346226} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.456498] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.456788] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.456988] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.457339] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 971.458029] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20dda8c1-d435-49a2-a27a-5dd58596dffe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.465200] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 971.465200] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523375d0-eb48-60e7-d46f-d11b8472bbad" [ 971.465200] env[65385]: _type = "Task" [ 971.465200] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.479251] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523375d0-eb48-60e7-d46f-d11b8472bbad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.506189] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.791734} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.506783] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.509041] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c198a44-1675-4212-a826-edc4fd8ab7dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.535431] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 167e1b46-af52-4a7a-9964-edabc8155287/167e1b46-af52-4a7a-9964-edabc8155287.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.537731] env[65385]: DEBUG nova.scheduler.client.report [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 971.542885] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1431a02-6890-4a07-8d1c-27fc8cf42b10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.565556] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 971.565556] env[65385]: value = "task-4454117" [ 971.565556] env[65385]: _type = "Task" [ 971.565556] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.575154] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.632475] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454116, 'name': CreateVM_Task, 'duration_secs': 0.382694} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.632965] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.633257] env[65385]: WARNING neutronclient.v2_0.client [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 971.633631] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.637234] env[65385]: WARNING neutronclient.v2_0.client [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 971.637876] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.638263] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.727922] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.755131] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.755599] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.822509] env[65385]: WARNING neutronclient.v2_0.client [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 971.823165] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.823523] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.889931] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.901311] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454115, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.912674] env[65385]: DEBUG nova.network.neutron [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 971.939461] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d8919ce7-7ae7-405e-b451-caaf5efc4148 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.737s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.976238] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523375d0-eb48-60e7-d46f-d11b8472bbad, 'name': SearchDatastore_Task, 'duration_secs': 0.024559} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.976545] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.976769] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.977028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.977198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.977338] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.977624] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.977925] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 971.978185] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58c7c2b2-7b32-4c0f-89a2-1a947ba693ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.980150] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5b66759-0664-46bc-9701-76c16eff63b3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.985900] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 971.985900] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5298e3ea-e49c-ece3-be35-8dd976a27d97" [ 971.985900] env[65385]: _type = "Task" [ 971.985900] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.990902] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.991173] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 971.992336] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-495097b5-5648-41d9-bcad-760ef05ae296 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.997757] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5298e3ea-e49c-ece3-be35-8dd976a27d97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.001253] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 972.001253] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528dbdee-0317-9b0f-0400-241bbacc7b96" [ 972.001253] env[65385]: _type = "Task" [ 972.001253] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.009960] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528dbdee-0317-9b0f-0400-241bbacc7b96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.044012] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.046840] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.072s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.047101] env[65385]: DEBUG nova.objects.instance [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lazy-loading 'resources' on Instance uuid 354527f0-007d-449f-9e15-48ce1d91876f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.072009] env[65385]: INFO nova.scheduler.client.report [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted allocations for instance 4b0b8859-b63e-4740-bf67-5733c75e1c70 [ 972.080381] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454117, 'name': ReconfigVM_Task, 'duration_secs': 0.305294} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.080691] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 167e1b46-af52-4a7a-9964-edabc8155287/167e1b46-af52-4a7a-9964-edabc8155287.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.081376] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67da835f-1912-4153-ab79-b4a50dc517f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.091991] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 972.091991] env[65385]: value = "task-4454118" [ 972.091991] env[65385]: _type = "Task" [ 972.091991] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.102281] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454118, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.400921] env[65385]: DEBUG oslo_vmware.api [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454115, 'name': RemoveSnapshot_Task, 'duration_secs': 1.228098} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.401211] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 972.416890] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 972.417311] env[65385]: WARNING neutronclient.v2_0.client [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.418058] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 972.418412] env[65385]: WARNING openstack [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 972.423702] env[65385]: WARNING neutronclient.v2_0.client [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.451162] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3ac044-8592-4d6a-833e-046b47012071 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.461908] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f023a9c4-57dc-45e6-bf17-a8609bd9f44c tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Suspending the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 972.462169] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6c7a5878-d81d-4287-ab42-1647b49d1cd0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.471116] env[65385]: DEBUG oslo_vmware.api [None req-f023a9c4-57dc-45e6-bf17-a8609bd9f44c tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 972.471116] env[65385]: value = "task-4454119" [ 972.471116] env[65385]: _type = "Task" [ 972.471116] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.481272] env[65385]: DEBUG oslo_vmware.api [None req-f023a9c4-57dc-45e6-bf17-a8609bd9f44c tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454119, 'name': SuspendVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.499760] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5298e3ea-e49c-ece3-be35-8dd976a27d97, 'name': SearchDatastore_Task, 'duration_secs': 0.011145} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.500633] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 972.500633] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.500815] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.518348] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528dbdee-0317-9b0f-0400-241bbacc7b96, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.519539] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e7f04f-6340-45a0-9c89-f4f56e187bbc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.526924] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 972.526924] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9050d-1d34-5962-954c-0e547182a9d4" [ 972.526924] env[65385]: _type = "Task" [ 972.526924] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.537848] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9050d-1d34-5962-954c-0e547182a9d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.582761] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a374032b-d244-4a18-8b74-f859779babc2 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "4b0b8859-b63e-4740-bf67-5733c75e1c70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.776s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.607404] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454118, 'name': Rename_Task, 'duration_secs': 0.154385} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.607404] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.609447] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c59fb6f5-9d1f-40ca-bd90-9efd3c49dda3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.617540] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 972.617540] env[65385]: value = "task-4454120" [ 972.617540] env[65385]: _type = "Task" [ 972.617540] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.630853] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454120, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.763159] env[65385]: DEBUG nova.compute.manager [req-03a2310d-b21f-45ae-bf06-2c5c22ab6336 req-4b85c9b0-6574-44d2-9eed-374e153c1b3f service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-vif-unplugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 972.763383] env[65385]: DEBUG oslo_concurrency.lockutils [req-03a2310d-b21f-45ae-bf06-2c5c22ab6336 req-4b85c9b0-6574-44d2-9eed-374e153c1b3f service nova] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.763615] env[65385]: DEBUG oslo_concurrency.lockutils [req-03a2310d-b21f-45ae-bf06-2c5c22ab6336 req-4b85c9b0-6574-44d2-9eed-374e153c1b3f service nova] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.765114] env[65385]: DEBUG oslo_concurrency.lockutils [req-03a2310d-b21f-45ae-bf06-2c5c22ab6336 req-4b85c9b0-6574-44d2-9eed-374e153c1b3f service nova] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.765347] env[65385]: DEBUG nova.compute.manager [req-03a2310d-b21f-45ae-bf06-2c5c22ab6336 req-4b85c9b0-6574-44d2-9eed-374e153c1b3f service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] No waiting events found dispatching network-vif-unplugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 972.765560] env[65385]: WARNING nova.compute.manager [req-03a2310d-b21f-45ae-bf06-2c5c22ab6336 req-4b85c9b0-6574-44d2-9eed-374e153c1b3f service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received unexpected event network-vif-unplugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 for instance with vm_state shelved and task_state shelving_offloading. [ 972.864028] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4721a0ed-9dcf-4391-815a-8d36a21a9e2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.872315] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb881c5-dbf6-4fe6-8d6c-eb3f541bade8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.910238] env[65385]: WARNING nova.compute.manager [None req-01feff41-774b-4986-8e66-87bed6f97acc tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Image not found during snapshot: nova.exception.ImageNotFound: Image 034b156e-8cd8-44b1-b387-b759f79925e6 could not be found. [ 972.912802] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.913821] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3132d017-989f-4515-a12b-6fc4d5c77691 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.917205] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfe586e-afa9-487e-81fa-4f494d531675 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.929663] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.931356] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fbaa11-fe0b-4110-b1cd-3b2641f56fe0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.937331] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-172769d3-7dce-402f-9852-232b46dab821 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.953083] env[65385]: DEBUG nova.compute.provider_tree [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.982567] env[65385]: DEBUG oslo_vmware.api [None req-f023a9c4-57dc-45e6-bf17-a8609bd9f44c tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454119, 'name': SuspendVM_Task} progress is 62%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.009525] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 973.009714] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 973.009944] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleting the datastore file [datastore2] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.010301] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f01f3975-c4f4-4c93-8665-d2dd6910e4ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.018091] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 973.018091] env[65385]: value = "task-4454122" [ 973.018091] env[65385]: _type = "Task" [ 973.018091] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.027245] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.036331] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9050d-1d34-5962-954c-0e547182a9d4, 'name': SearchDatastore_Task, 'duration_secs': 0.021275} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.036587] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 973.036838] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.037124] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 973.037306] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 973.037610] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28ad6206-c86d-4c36-83f3-7cf565bf6d30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.039760] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6dfcb1c3-bcd0-449e-8148-3da1bb267c86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.048809] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 973.048809] env[65385]: value = "task-4454123" [ 973.048809] env[65385]: _type = "Task" [ 973.048809] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.057781] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.060743] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 973.060916] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 973.061668] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17d16c4a-7c7f-4d9d-b246-8ce7fb7c8509 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.067765] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 973.067765] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f36dbe-b11d-7972-068e-f1515e2faa52" [ 973.067765] env[65385]: _type = "Task" [ 973.067765] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.077143] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f36dbe-b11d-7972-068e-f1515e2faa52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.131655] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454120, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.456535] env[65385]: DEBUG nova.scheduler.client.report [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 973.489710] env[65385]: DEBUG oslo_vmware.api [None req-f023a9c4-57dc-45e6-bf17-a8609bd9f44c tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454119, 'name': SuspendVM_Task, 'duration_secs': 0.678158} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.490641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f023a9c4-57dc-45e6-bf17-a8609bd9f44c tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Suspended the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 973.490641] env[65385]: DEBUG nova.compute.manager [None req-f023a9c4-57dc-45e6-bf17-a8609bd9f44c tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 973.491729] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62a36ef-916f-49ba-b1b6-7e200bd7d22a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.536930] env[65385]: DEBUG oslo_vmware.api [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.235373} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.537344] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.537716] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.538072] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.562874] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454123, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.576181] env[65385]: INFO nova.scheduler.client.report [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted allocations for instance 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 [ 973.584519] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f36dbe-b11d-7972-068e-f1515e2faa52, 'name': SearchDatastore_Task, 'duration_secs': 0.010494} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.585278] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c836cce8-739a-436a-9d31-60823d7efedc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.591871] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 973.591871] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527e0fd5-c81f-47d4-ab6c-f12cc53ee8e5" [ 973.591871] env[65385]: _type = "Task" [ 973.591871] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.601308] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527e0fd5-c81f-47d4-ab6c-f12cc53ee8e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.632311] env[65385]: DEBUG oslo_vmware.api [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454120, 'name': PowerOnVM_Task, 'duration_secs': 0.719848} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.632626] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.632854] env[65385]: INFO nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Took 7.82 seconds to spawn the instance on the hypervisor. [ 973.633226] env[65385]: DEBUG nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 973.634131] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e542dcd3-205f-446e-afc1-6e9d73e701cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.871924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "3febf6af-699a-4fa1-b079-3790cd1095e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.872186] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.872434] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "3febf6af-699a-4fa1-b079-3790cd1095e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.872629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.872798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.874826] env[65385]: INFO nova.compute.manager [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Terminating instance [ 973.963550] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.916s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.965798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.673s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.966009] env[65385]: DEBUG nova.objects.instance [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lazy-loading 'resources' on Instance uuid f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.984520] env[65385]: INFO nova.scheduler.client.report [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Deleted allocations for instance 354527f0-007d-449f-9e15-48ce1d91876f [ 974.060864] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454123, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.085630] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.103341] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527e0fd5-c81f-47d4-ab6c-f12cc53ee8e5, 'name': SearchDatastore_Task, 'duration_secs': 0.019598} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.103643] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.103921] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 974.104204] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3509029a-ab24-4cdc-a57f-6a37cae780f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.112870] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 974.112870] env[65385]: value = "task-4454124" [ 974.112870] env[65385]: _type = "Task" [ 974.112870] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.122288] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.153877] env[65385]: INFO nova.compute.manager [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Took 24.60 seconds to build instance. [ 974.167167] env[65385]: DEBUG oslo_concurrency.lockutils [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "26a2d464-0aab-463f-8265-948a4f0ff188" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.167508] env[65385]: DEBUG oslo_concurrency.lockutils [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.336863] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "f4af288e-b9c9-4273-8e65-e726ae039d03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.337197] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.337445] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "f4af288e-b9c9-4273-8e65-e726ae039d03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.337636] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.337840] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.339967] env[65385]: INFO nova.compute.manager [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Terminating instance [ 974.372315] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "interface-167e1b46-af52-4a7a-9964-edabc8155287-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.372628] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "interface-167e1b46-af52-4a7a-9964-edabc8155287-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.373084] env[65385]: DEBUG nova.objects.instance [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lazy-loading 'flavor' on Instance uuid 167e1b46-af52-4a7a-9964-edabc8155287 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.378858] env[65385]: DEBUG nova.compute.manager [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 974.379227] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.379945] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e7a1c5-3fbf-495a-8b3d-54684c7c6803 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.389117] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.389362] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cfb966c6-67bf-4ce8-a7c4-920574e9d2bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.397660] env[65385]: DEBUG oslo_vmware.api [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 974.397660] env[65385]: value = "task-4454125" [ 974.397660] env[65385]: _type = "Task" [ 974.397660] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.406644] env[65385]: DEBUG oslo_vmware.api [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4454125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.493999] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1534a9f-0302-47c9-9214-73cc8ece7402 tempest-ServersListShow298Test-1198603314 tempest-ServersListShow298Test-1198603314-project-member] Lock "354527f0-007d-449f-9e15-48ce1d91876f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.054s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.563724] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454123, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.623446] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.656597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5a576c2d-7247-4b14-85fa-1000617e0089 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "167e1b46-af52-4a7a-9964-edabc8155287" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.113s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.668054] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcdf4c5-f1f1-43bf-bc27-ef4d22a7898c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.673223] env[65385]: INFO nova.compute.manager [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Detaching volume 33fd2b93-a90e-4837-9089-e9d5cab1e2df [ 974.684874] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b567057f-f120-4cb6-9ec3-89e6b8de1b3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.727784] env[65385]: INFO nova.virt.block_device [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Attempting to driver detach volume 33fd2b93-a90e-4837-9089-e9d5cab1e2df from mountpoint /dev/sdb [ 974.728360] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 974.728360] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871070', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'name': 'volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '26a2d464-0aab-463f-8265-948a4f0ff188', 'attached_at': '', 'detached_at': '', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'serial': '33fd2b93-a90e-4837-9089-e9d5cab1e2df'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 974.729052] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e257b7-a3de-4038-944a-41d71234b74b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.734244] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1412894-7adc-4d5e-99ea-16a40bdb6b36 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.761137] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c35a85c-27e5-447a-86e2-fe771a20606d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.765875] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff9af59-4783-4e31-a9e6-77787605d5ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.783587] env[65385]: DEBUG nova.compute.provider_tree [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.786263] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a9c2d3-d7f2-4957-8fc5-272aa843d027 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.814599] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3027b8ea-8630-442c-8e43-8a479ec104c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.837301] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The volume has not been displaced from its original location: [datastore2] volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df/volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 974.839925] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Reconfiguring VM instance instance-00000037 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 974.840485] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c56a770-08ca-4b15-86fd-f205b75bcc0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.854496] env[65385]: DEBUG nova.compute.manager [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 974.854756] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.858460] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e98b34-bfb9-494e-99b9-3e02796944fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.867306] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.868894] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0def62e-406d-4e28-adbd-ab4f7cc263c5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.870758] env[65385]: DEBUG oslo_vmware.api [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 974.870758] env[65385]: value = "task-4454126" [ 974.870758] env[65385]: _type = "Task" [ 974.870758] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.876707] env[65385]: WARNING neutronclient.v2_0.client [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 974.877419] env[65385]: WARNING openstack [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 974.877855] env[65385]: WARNING openstack [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 974.883633] env[65385]: DEBUG nova.objects.instance [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lazy-loading 'pci_requests' on Instance uuid 167e1b46-af52-4a7a-9964-edabc8155287 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.890547] env[65385]: DEBUG oslo_vmware.api [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 974.890547] env[65385]: value = "task-4454127" [ 974.890547] env[65385]: _type = "Task" [ 974.890547] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.902241] env[65385]: DEBUG oslo_vmware.api [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.913214] env[65385]: DEBUG oslo_vmware.api [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4454125, 'name': PowerOffVM_Task, 'duration_secs': 0.454081} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.913476] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.913634] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.913890] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1d02f4f-7b74-40d9-872f-b394739d655d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.036742] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 975.037263] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 975.037323] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Deleting the datastore file [datastore2] 3febf6af-699a-4fa1-b079-3790cd1095e5 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.037608] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1cf045b-54d8-44d7-9970-dfd80bd44368 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.050491] env[65385]: DEBUG oslo_vmware.api [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for the task: (returnval){ [ 975.050491] env[65385]: value = "task-4454129" [ 975.050491] env[65385]: _type = "Task" [ 975.050491] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.065377] env[65385]: DEBUG oslo_vmware.api [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4454129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.069924] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454123, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.593273} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.070273] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 975.070535] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.070872] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-028dddea-afe4-4e51-babf-f538b8990ae3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.081082] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 975.081082] env[65385]: value = "task-4454130" [ 975.081082] env[65385]: _type = "Task" [ 975.081082] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.091577] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.127188] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454124, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.290365] env[65385]: DEBUG nova.scheduler.client.report [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 975.356059] env[65385]: DEBUG nova.compute.manager [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 975.356264] env[65385]: DEBUG nova.compute.manager [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing instance network info cache due to event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 975.356474] env[65385]: DEBUG oslo_concurrency.lockutils [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Acquiring lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.356612] env[65385]: DEBUG oslo_concurrency.lockutils [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Acquired lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.356763] env[65385]: DEBUG nova.network.neutron [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 975.382165] env[65385]: DEBUG oslo_vmware.api [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454126, 'name': ReconfigVM_Task, 'duration_secs': 0.46965} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.382501] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Reconfigured VM instance instance-00000037 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 975.387573] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fcc78527-4b05-44b5-a545-18f1f673a6be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.398291] env[65385]: DEBUG nova.objects.base [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Object Instance<167e1b46-af52-4a7a-9964-edabc8155287> lazy-loaded attributes: flavor,pci_requests {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 975.398503] env[65385]: DEBUG nova.network.neutron [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 975.398837] env[65385]: WARNING neutronclient.v2_0.client [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.399150] env[65385]: WARNING neutronclient.v2_0.client [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.399730] env[65385]: WARNING openstack [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.400091] env[65385]: WARNING openstack [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.418337] env[65385]: DEBUG oslo_vmware.api [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454127, 'name': PowerOffVM_Task, 'duration_secs': 0.402842} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.419054] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 975.419232] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 975.419551] env[65385]: DEBUG oslo_vmware.api [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 975.419551] env[65385]: value = "task-4454131" [ 975.419551] env[65385]: _type = "Task" [ 975.419551] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.419763] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b920cf1-837b-41de-845e-52d8539b411c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.433624] env[65385]: DEBUG oslo_vmware.api [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454131, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.498120] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 975.498504] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 975.499667] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleting the datastore file [datastore1] f4af288e-b9c9-4273-8e65-e726ae039d03 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.499667] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-858fc89c-f010-4893-a0a1-07efa73b2fde {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.510332] env[65385]: DEBUG oslo_vmware.api [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for the task: (returnval){ [ 975.510332] env[65385]: value = "task-4454133" [ 975.510332] env[65385]: _type = "Task" [ 975.510332] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.521097] env[65385]: DEBUG oslo_vmware.api [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.538329] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a05003e-ad39-4062-af24-f9c540878bb1 tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "interface-167e1b46-af52-4a7a-9964-edabc8155287-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.165s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.563216] env[65385]: DEBUG oslo_vmware.api [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Task: {'id': task-4454129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234314} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.563980] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.564238] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.564511] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.564729] env[65385]: INFO nova.compute.manager [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Took 1.19 seconds to destroy the instance on the hypervisor. [ 975.565045] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 975.565382] env[65385]: DEBUG nova.compute.manager [-] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 975.565524] env[65385]: DEBUG nova.network.neutron [-] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 975.565872] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.566596] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.566870] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.591483] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090001} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.591818] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.592686] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00892b96-827c-48c6-a589-6f89190d4b40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.615188] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.615580] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68c68d81-3aa5-4f3c-9a36-695301639fd7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.630908] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.642563] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454124, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.044472} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.643918] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 975.644154] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.644464] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 975.644464] env[65385]: value = "task-4454134" [ 975.644464] env[65385]: _type = "Task" [ 975.644464] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.644681] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78f0a879-14da-4cf8-bc68-714d898bd241 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.655791] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454134, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.657323] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 975.657323] env[65385]: value = "task-4454135" [ 975.657323] env[65385]: _type = "Task" [ 975.657323] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.670088] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454135, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.695183] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 975.695650] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.695986] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 975.696307] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.696597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.699368] env[65385]: INFO nova.compute.manager [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Terminating instance [ 975.796853] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.799756] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.210s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.800033] env[65385]: DEBUG nova.objects.instance [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lazy-loading 'resources' on Instance uuid 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.818725] env[65385]: INFO nova.scheduler.client.report [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted allocations for instance f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5 [ 975.859944] env[65385]: WARNING neutronclient.v2_0.client [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.860763] env[65385]: WARNING openstack [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 975.861156] env[65385]: WARNING openstack [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 975.936034] env[65385]: DEBUG oslo_vmware.api [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454131, 'name': ReconfigVM_Task, 'duration_secs': 0.172574} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.936034] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871070', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'name': 'volume-33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '26a2d464-0aab-463f-8265-948a4f0ff188', 'attached_at': '', 'detached_at': '', 'volume_id': '33fd2b93-a90e-4837-9089-e9d5cab1e2df', 'serial': '33fd2b93-a90e-4837-9089-e9d5cab1e2df'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 976.025042] env[65385]: DEBUG oslo_vmware.api [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Task: {'id': task-4454133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152906} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.025750] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.025750] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.026574] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.026574] env[65385]: INFO nova.compute.manager [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Took 1.17 seconds to destroy the instance on the hypervisor. [ 976.026574] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 976.026574] env[65385]: DEBUG nova.compute.manager [-] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 976.026574] env[65385]: DEBUG nova.network.neutron [-] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 976.026787] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 976.027414] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.027491] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.108191] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 976.159810] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454134, 'name': ReconfigVM_Task, 'duration_secs': 0.300711} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.163786] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.165235] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-696d8c97-3739-4c27-a2ff-5b3dd1f059aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.173862] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454135, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126098} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.176691] env[65385]: WARNING openstack [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.177433] env[65385]: WARNING openstack [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.185125] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.186142] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 976.186142] env[65385]: value = "task-4454136" [ 976.186142] env[65385]: _type = "Task" [ 976.186142] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.187744] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92e10d5-03f4-46eb-a600-362e8589decf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.211313] env[65385]: DEBUG nova.compute.manager [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 976.211566] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 976.220656] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.225447] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26896e57-3fde-4fbc-a9fa-cef3938ba154 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.228889] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-699b7d7d-0ebc-4090-964f-b8f285f85e5b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.251984] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 976.253445] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db8a091f-8b87-4af0-a2aa-a415b0d3fdac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.254975] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 976.254975] env[65385]: value = "task-4454137" [ 976.254975] env[65385]: _type = "Task" [ 976.254975] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.265113] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454137, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.327780] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4cc33f1e-af57-4a51-9a56-678c2ece06b6 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.609s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.335945] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 976.336270] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 976.336574] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleting the datastore file [datastore2] c003dd80-d419-4a5a-8f86-dcac1a1e789c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.336894] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a817342b-c553-4532-a415-c8fd7216757d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.345185] env[65385]: DEBUG oslo_vmware.api [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 976.345185] env[65385]: value = "task-4454139" [ 976.345185] env[65385]: _type = "Task" [ 976.345185] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.360088] env[65385]: DEBUG oslo_vmware.api [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.473695] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.479047] env[65385]: WARNING neutronclient.v2_0.client [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 976.479047] env[65385]: WARNING openstack [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.479047] env[65385]: WARNING openstack [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.539630] env[65385]: DEBUG nova.objects.instance [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'flavor' on Instance uuid 26a2d464-0aab-463f-8265-948a4f0ff188 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.596665] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0655584-b9f0-4ad2-9310-1e6729c2b0d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.606841] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8f8815-d65b-4091-a4f4-0453fc3cd075 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.643705] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b04675-a515-4f3c-8592-36d62975e787 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.651098] env[65385]: DEBUG nova.network.neutron [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updated VIF entry in instance network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 976.651098] env[65385]: DEBUG nova.network.neutron [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 976.655615] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84009a77-3162-4bad-bcce-ce4f0c206a0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.672877] env[65385]: DEBUG nova.compute.provider_tree [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.704181] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454136, 'name': Rename_Task, 'duration_secs': 0.180985} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.704268] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.706486] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7813a9cf-fe4d-45a2-b7e6-008a98f0dc9e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.716399] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 976.716399] env[65385]: value = "task-4454140" [ 976.716399] env[65385]: _type = "Task" [ 976.716399] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.728534] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.769386] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454137, 'name': ReconfigVM_Task, 'duration_secs': 0.288344} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.769386] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55/370b6aa8-2337-48e1-8102-611ce4494a55.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.769906] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f1158d0-37f7-4160-a661-278b79f92f33 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.777202] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 976.777202] env[65385]: value = "task-4454141" [ 976.777202] env[65385]: _type = "Task" [ 976.777202] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.787259] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454141, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.860056] env[65385]: DEBUG oslo_vmware.api [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170912} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.860056] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.860056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.860732] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.860732] env[65385]: INFO nova.compute.manager [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Took 0.65 seconds to destroy the instance on the hypervisor. [ 976.860732] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 976.861027] env[65385]: DEBUG nova.compute.manager [-] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 976.861027] env[65385]: DEBUG nova.network.neutron [-] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 976.861236] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 976.861827] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.862293] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.897123] env[65385]: DEBUG nova.network.neutron [-] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 976.907581] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 977.019942] env[65385]: DEBUG nova.compute.manager [req-44b8722e-ba14-41ba-8c38-6640b52c0624 req-c1d6888c-fea1-45b9-84dd-0a56a9c12d8d service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Received event network-vif-deleted-37cdfc1d-4b24-4ae6-ac71-e0d63602b872 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 977.020069] env[65385]: INFO nova.compute.manager [req-44b8722e-ba14-41ba-8c38-6640b52c0624 req-c1d6888c-fea1-45b9-84dd-0a56a9c12d8d service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Neutron deleted interface 37cdfc1d-4b24-4ae6-ac71-e0d63602b872; detaching it from the instance and deleting it from the info cache [ 977.020234] env[65385]: DEBUG nova.network.neutron [req-44b8722e-ba14-41ba-8c38-6640b52c0624 req-c1d6888c-fea1-45b9-84dd-0a56a9c12d8d service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 977.155753] env[65385]: DEBUG oslo_concurrency.lockutils [req-47010008-bc1a-4e46-8d57-f028ea8b8be7 req-ea59abf9-bac1-4cd3-989b-3d688b5d1f37 service nova] Releasing lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 977.177129] env[65385]: DEBUG nova.scheduler.client.report [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.210034] env[65385]: DEBUG nova.network.neutron [-] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 977.229202] env[65385]: DEBUG oslo_vmware.api [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454140, 'name': PowerOnVM_Task, 'duration_secs': 0.503958} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.229372] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 977.229575] env[65385]: INFO nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Took 7.42 seconds to spawn the instance on the hypervisor. [ 977.229751] env[65385]: DEBUG nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 977.230577] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fbabfc-8a7c-4ff9-8501-5f24e0f5a0ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.289655] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454141, 'name': Rename_Task, 'duration_secs': 0.175675} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.290081] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.290368] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e5987b1-4e89-4c24-8d98-a21df9ced85a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.298773] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 977.298773] env[65385]: value = "task-4454142" [ 977.298773] env[65385]: _type = "Task" [ 977.298773] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.309623] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454142, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.391973] env[65385]: DEBUG nova.compute.manager [req-eb573a4f-dd0c-44f1-b1c4-c6bf7a6559e0 req-3d9c5b7c-9a70-4d5a-9f4c-5537a1a6342d service nova] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Received event network-vif-deleted-2a55996b-b9ef-4e79-9979-f42e6bdd2127 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 977.391973] env[65385]: DEBUG nova.compute.manager [req-eb573a4f-dd0c-44f1-b1c4-c6bf7a6559e0 req-3d9c5b7c-9a70-4d5a-9f4c-5537a1a6342d service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Received event network-vif-deleted-82a01c37-b04c-4b2c-a613-dc4117b64ef2 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 977.392323] env[65385]: INFO nova.compute.manager [req-eb573a4f-dd0c-44f1-b1c4-c6bf7a6559e0 req-3d9c5b7c-9a70-4d5a-9f4c-5537a1a6342d service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Neutron deleted interface 82a01c37-b04c-4b2c-a613-dc4117b64ef2; detaching it from the instance and deleting it from the info cache [ 977.392421] env[65385]: DEBUG nova.network.neutron [req-eb573a4f-dd0c-44f1-b1c4-c6bf7a6559e0 req-3d9c5b7c-9a70-4d5a-9f4c-5537a1a6342d service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 977.400122] env[65385]: INFO nova.compute.manager [-] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Took 1.37 seconds to deallocate network for instance. [ 977.523352] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f996ffcf-c6d6-4d56-b99a-5d832873ac00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.534097] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9728080-b6b5-4432-9910-08279919b169 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.549124] env[65385]: DEBUG oslo_concurrency.lockutils [None req-27b04dbd-b296-41c1-b8af-fa813abceaa7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.382s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.567224] env[65385]: DEBUG nova.compute.manager [req-44b8722e-ba14-41ba-8c38-6640b52c0624 req-c1d6888c-fea1-45b9-84dd-0a56a9c12d8d service nova] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Detach interface failed, port_id=37cdfc1d-4b24-4ae6-ac71-e0d63602b872, reason: Instance 3febf6af-699a-4fa1-b079-3790cd1095e5 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 977.609922] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "167e1b46-af52-4a7a-9964-edabc8155287" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.610467] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "167e1b46-af52-4a7a-9964-edabc8155287" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.610529] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "167e1b46-af52-4a7a-9964-edabc8155287-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.610680] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "167e1b46-af52-4a7a-9964-edabc8155287-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.610842] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "167e1b46-af52-4a7a-9964-edabc8155287-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.612797] env[65385]: INFO nova.compute.manager [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Terminating instance [ 977.626713] env[65385]: DEBUG nova.network.neutron [-] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 977.685011] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.688057] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.428s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.689449] env[65385]: INFO nova.compute.claims [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.710631] env[65385]: INFO nova.scheduler.client.report [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted allocations for instance 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3 [ 977.715116] env[65385]: INFO nova.compute.manager [-] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Took 2.15 seconds to deallocate network for instance. [ 977.751858] env[65385]: INFO nova.compute.manager [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Took 22.85 seconds to build instance. [ 977.810709] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454142, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.895096] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c07f400b-1cb3-4a11-a856-42f390a27fe0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.905614] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc2bb21-e0ee-4cd1-8b69-95a8746e7144 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.918049] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.938052] env[65385]: DEBUG nova.compute.manager [req-eb573a4f-dd0c-44f1-b1c4-c6bf7a6559e0 req-3d9c5b7c-9a70-4d5a-9f4c-5537a1a6342d service nova] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Detach interface failed, port_id=82a01c37-b04c-4b2c-a613-dc4117b64ef2, reason: Instance c003dd80-d419-4a5a-8f86-dcac1a1e789c could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 978.116379] env[65385]: DEBUG nova.compute.manager [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 978.116636] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.117927] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ec1c73-b255-463d-bfc5-d057c91bdc02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.127359] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.127949] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2eee2fff-7f0c-4e85-9acb-d9995fe11c23 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.129814] env[65385]: INFO nova.compute.manager [-] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Took 1.27 seconds to deallocate network for instance. [ 978.138213] env[65385]: DEBUG oslo_vmware.api [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 978.138213] env[65385]: value = "task-4454143" [ 978.138213] env[65385]: _type = "Task" [ 978.138213] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.148809] env[65385]: DEBUG oslo_vmware.api [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.218286] env[65385]: DEBUG oslo_concurrency.lockutils [None req-70746683-38c5-43be-8122-33b3cc546a05 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.420s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.220446] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.255315] env[65385]: DEBUG oslo_concurrency.lockutils [None req-926f42ba-53b4-4733-84bb-42d9d4e5bd77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.367s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.312028] env[65385]: DEBUG oslo_vmware.api [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454142, 'name': PowerOnVM_Task, 'duration_secs': 1.01133} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.312458] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.312747] env[65385]: DEBUG nova.compute.manager [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 978.313697] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedee241-8a55-4a71-9bf5-c0bf84bdc7f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.596676] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "26a2d464-0aab-463f-8265-948a4f0ff188" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.596955] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.597199] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "26a2d464-0aab-463f-8265-948a4f0ff188-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.597398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.597571] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.600197] env[65385]: INFO nova.compute.manager [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Terminating instance [ 978.638082] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.650114] env[65385]: DEBUG oslo_vmware.api [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454143, 'name': PowerOffVM_Task, 'duration_secs': 0.325164} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.650412] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.650575] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.650829] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fdf6bc3b-bfc4-420a-b0e8-1d9f51c99366 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.721985] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.722226] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.722457] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Deleting the datastore file [datastore2] 167e1b46-af52-4a7a-9964-edabc8155287 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.725305] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46eaf9f5-06bd-426d-98d3-922699f9c61e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.732706] env[65385]: DEBUG oslo_vmware.api [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for the task: (returnval){ [ 978.732706] env[65385]: value = "task-4454145" [ 978.732706] env[65385]: _type = "Task" [ 978.732706] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.742825] env[65385]: DEBUG oslo_vmware.api [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454145, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.770487] env[65385]: INFO nova.compute.manager [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Rebuilding instance [ 978.833361] env[65385]: DEBUG nova.compute.manager [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 978.833806] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83404b1c-678a-4587-a2c5-8c3037b3b4f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.839093] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.909687] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aed51bc-4116-4d76-8202-91d97c6e5697 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.919181] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d326dc20-fa64-48c8-9e73-06cddad0e884 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.950458] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226c7292-29e5-440f-bd24-9dea4de93cf0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.958320] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23100fd2-c4d3-4d89-a3f8-bb49997f22ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.973838] env[65385]: DEBUG nova.compute.provider_tree [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.104656] env[65385]: DEBUG nova.compute.manager [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 979.104907] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.105931] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7423d98-1b8d-492d-a2fb-2e664df33de7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.114500] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.114745] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75a218ff-e599-4081-9337-7f199e7a6955 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.121059] env[65385]: DEBUG oslo_vmware.api [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 979.121059] env[65385]: value = "task-4454146" [ 979.121059] env[65385]: _type = "Task" [ 979.121059] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.129252] env[65385]: DEBUG oslo_vmware.api [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.243712] env[65385]: DEBUG oslo_vmware.api [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Task: {'id': task-4454145, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150414} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.244043] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.244284] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.244362] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.244532] env[65385]: INFO nova.compute.manager [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Took 1.13 seconds to destroy the instance on the hypervisor. [ 979.244768] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 979.244972] env[65385]: DEBUG nova.compute.manager [-] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 979.245082] env[65385]: DEBUG nova.network.neutron [-] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 979.245328] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.245880] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 979.246149] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 979.289020] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.309621] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.309621] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.341812] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.342152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.478083] env[65385]: DEBUG nova.scheduler.client.report [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.608195] env[65385]: DEBUG nova.compute.manager [req-fd9a66d1-d694-4632-9e18-055389c8c954 req-87d90a1c-3a79-4d35-9a32-474944109962 service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Received event network-vif-deleted-06080c91-c437-4033-b17e-28344fc48b5a {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 979.608469] env[65385]: INFO nova.compute.manager [req-fd9a66d1-d694-4632-9e18-055389c8c954 req-87d90a1c-3a79-4d35-9a32-474944109962 service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Neutron deleted interface 06080c91-c437-4033-b17e-28344fc48b5a; detaching it from the instance and deleting it from the info cache [ 979.608568] env[65385]: DEBUG nova.network.neutron [req-fd9a66d1-d694-4632-9e18-055389c8c954 req-87d90a1c-3a79-4d35-9a32-474944109962 service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 979.632271] env[65385]: DEBUG oslo_vmware.api [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454146, 'name': PowerOffVM_Task, 'duration_secs': 0.26596} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.632412] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.632563] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.632826] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e81e874-9702-4d6b-a14f-39f28ffb5723 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.704294] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.704629] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.704859] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleting the datastore file [datastore1] 26a2d464-0aab-463f-8265-948a4f0ff188 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.705161] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cd425c4-caac-431b-94d2-88f23b450aef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.712464] env[65385]: DEBUG oslo_vmware.api [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 979.712464] env[65385]: value = "task-4454148" [ 979.712464] env[65385]: _type = "Task" [ 979.712464] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.722784] env[65385]: DEBUG oslo_vmware.api [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.814079] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 979.846615] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 979.853218] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.854299] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed7cc91c-075e-4947-b90d-6edd74d3a5d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.863971] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 979.863971] env[65385]: value = "task-4454149" [ 979.863971] env[65385]: _type = "Task" [ 979.863971] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.876966] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.985637] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.986184] env[65385]: DEBUG nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 979.989252] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.261s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.989428] env[65385]: DEBUG nova.objects.instance [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lazy-loading 'pci_requests' on Instance uuid 0c345794-036c-45d1-985e-cf5c116bd873 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.050450] env[65385]: DEBUG nova.network.neutron [-] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 980.111683] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9cd369a-22dd-4005-b0a4-f2b082c8a3c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.122653] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05324647-e34c-4bcc-b128-34dc5ee04de2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.153978] env[65385]: DEBUG nova.compute.manager [req-fd9a66d1-d694-4632-9e18-055389c8c954 req-87d90a1c-3a79-4d35-9a32-474944109962 service nova] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Detach interface failed, port_id=06080c91-c437-4033-b17e-28344fc48b5a, reason: Instance 167e1b46-af52-4a7a-9964-edabc8155287 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 980.223223] env[65385]: DEBUG oslo_vmware.api [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152082} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.223594] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.223785] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.223963] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.224151] env[65385]: INFO nova.compute.manager [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Took 1.12 seconds to destroy the instance on the hypervisor. [ 980.224395] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 980.224596] env[65385]: DEBUG nova.compute.manager [-] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 980.224700] env[65385]: DEBUG nova.network.neutron [-] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 980.225023] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 980.225577] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 980.225841] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 980.262637] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 980.337672] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.366136] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.375905] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454149, 'name': PowerOffVM_Task, 'duration_secs': 0.363584} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.376201] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 980.376433] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.377268] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913eee2f-73ec-4a6c-89d3-ae5458d1b50e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.384793] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 980.385050] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15c64596-eb0a-40e1-b8b2-086d68cf7b6f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.412844] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 980.413086] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 980.413269] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Deleting the datastore file [datastore1] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.413555] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60a17418-eee0-4dbe-8978-46531465ad15 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.421343] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 980.421343] env[65385]: value = "task-4454151" [ 980.421343] env[65385]: _type = "Task" [ 980.421343] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.430010] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.492901] env[65385]: DEBUG nova.compute.utils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 980.496526] env[65385]: DEBUG nova.objects.instance [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lazy-loading 'numa_topology' on Instance uuid 0c345794-036c-45d1-985e-cf5c116bd873 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.497580] env[65385]: DEBUG nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 980.497580] env[65385]: DEBUG nova.network.neutron [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 980.497891] env[65385]: WARNING neutronclient.v2_0.client [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 980.498219] env[65385]: WARNING neutronclient.v2_0.client [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 980.499667] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 980.499667] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 980.510186] env[65385]: DEBUG nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 980.513544] env[65385]: INFO nova.compute.claims [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.553105] env[65385]: INFO nova.compute.manager [-] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Took 1.31 seconds to deallocate network for instance. [ 980.561195] env[65385]: DEBUG nova.policy [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 980.923369] env[65385]: DEBUG nova.network.neutron [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Successfully created port: 14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 980.940285] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167867} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.940285] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.940985] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.942042] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.066011] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.227922] env[65385]: DEBUG nova.network.neutron [-] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 981.525838] env[65385]: DEBUG nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 981.556169] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 981.556421] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 981.556575] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 981.556753] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 981.556892] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 981.557045] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 981.557284] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.557532] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 981.557704] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 981.557862] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 981.558041] env[65385]: DEBUG nova.virt.hardware [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 981.558895] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3be17a-bce9-4a27-949a-6d93e29c43a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.567540] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90604fa2-9325-43ca-9c2c-fbf59591fe32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.642271] env[65385]: DEBUG nova.compute.manager [req-7cabc928-90d0-42c3-b7a3-1527f09544f1 req-cec944ef-0f93-49fb-8bb9-cbc5d24c1d81 service nova] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Received event network-vif-deleted-622e8c02-7c83-4b27-9884-f6c399712a84 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 981.724832] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73108495-06ba-4f4a-a5d1-49a082bef90f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.730174] env[65385]: INFO nova.compute.manager [-] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Took 1.51 seconds to deallocate network for instance. [ 981.734543] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e0f838-db65-4308-80f1-6382d85801d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.771202] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c64fafb-6ec5-4629-9616-5399da0b5a62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.780596] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bab8b3-a17d-4f51-a532-10b8547ba7fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.795257] env[65385]: DEBUG nova.compute.provider_tree [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.978596] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 981.978882] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 981.979080] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 981.979279] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 981.979490] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 981.979677] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 981.979900] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.980072] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 981.980238] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 981.980401] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 981.980568] env[65385]: DEBUG nova.virt.hardware [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 981.981438] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed6d5b2-379d-4e69-b3e7-537bd464eb94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.990303] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387d52f2-70fb-4af2-ad94-3defda65d915 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.005617] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.011291] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 982.011591] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.011859] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-168f4329-e6ba-492c-9e3b-bf269f8f63e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.029029] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.029029] env[65385]: value = "task-4454152" [ 982.029029] env[65385]: _type = "Task" [ 982.029029] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.037346] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454152, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.243087] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.298739] env[65385]: DEBUG nova.scheduler.client.report [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 982.443369] env[65385]: DEBUG nova.network.neutron [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Successfully updated port: 14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 982.539012] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454152, 'name': CreateVM_Task, 'duration_secs': 0.342854} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.539381] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.539600] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.539760] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 982.540116] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 982.540382] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7f54c64-ce48-4406-bdd4-ed15e744a7d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.545407] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 982.545407] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c88b19-78cb-1333-63e8-fedcee576247" [ 982.545407] env[65385]: _type = "Task" [ 982.545407] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.553338] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c88b19-78cb-1333-63e8-fedcee576247, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.803796] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.814s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.804190] env[65385]: WARNING neutronclient.v2_0.client [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 982.806913] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.917s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.807159] env[65385]: DEBUG nova.objects.instance [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lazy-loading 'resources' on Instance uuid 55936da6-fcdf-4291-9b3a-23fa8b7a56ef {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.847069] env[65385]: INFO nova.network.neutron [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating port bbadc19c-78ee-4246-a53d-77f01c5ccea8 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 982.946067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.946067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 982.946067] env[65385]: DEBUG nova.network.neutron [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 983.059650] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c88b19-78cb-1333-63e8-fedcee576247, 'name': SearchDatastore_Task, 'duration_secs': 0.010373} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.060139] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 983.060662] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.060982] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.061161] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 983.061354] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.061647] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8702c7d-9b93-4388-96f7-b675327c8423 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.072073] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.072188] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.072948] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2823ba17-41a0-4410-bdae-d4ee538b272d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.079478] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 983.079478] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c86f8c-0ad7-a685-51eb-76bf9553e13e" [ 983.079478] env[65385]: _type = "Task" [ 983.079478] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.089759] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c86f8c-0ad7-a685-51eb-76bf9553e13e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.449129] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.449511] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.489532] env[65385]: DEBUG nova.network.neutron [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 983.527592] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df5b061-7880-478d-9e24-bc5634534047 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.535610] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9b0669-e42e-4380-9eea-0d28ecac61f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.566794] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996365e1-b947-412a-ad6d-54eea7cc254c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.570694] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.571068] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.587121] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844883b1-919f-432e-820f-30f2f8d885fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.594979] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c86f8c-0ad7-a685-51eb-76bf9553e13e, 'name': SearchDatastore_Task, 'duration_secs': 0.009747} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.595874] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba93787b-d528-4a52-b03d-081f84f8f6ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.606276] env[65385]: DEBUG nova.compute.provider_tree [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.612865] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 983.612865] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b87b1e-afe6-b655-fb17-6761f0c7c381" [ 983.612865] env[65385]: _type = "Task" [ 983.612865] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.629759] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b87b1e-afe6-b655-fb17-6761f0c7c381, 'name': SearchDatastore_Task, 'duration_secs': 0.0103} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.630101] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 983.630369] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 983.630630] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac721bce-2d9d-4616-9284-06e8faf4e253 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.638046] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 983.638046] env[65385]: value = "task-4454153" [ 983.638046] env[65385]: _type = "Task" [ 983.638046] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.647090] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.661592] env[65385]: WARNING neutronclient.v2_0.client [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.662457] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.662733] env[65385]: WARNING openstack [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.684252] env[65385]: DEBUG nova.compute.manager [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-vif-plugged-14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 983.684355] env[65385]: DEBUG oslo_concurrency.lockutils [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.684553] env[65385]: DEBUG oslo_concurrency.lockutils [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.684727] env[65385]: DEBUG oslo_concurrency.lockutils [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.684912] env[65385]: DEBUG nova.compute.manager [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] No waiting events found dispatching network-vif-plugged-14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 983.685116] env[65385]: WARNING nova.compute.manager [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received unexpected event network-vif-plugged-14fd5163-1262-4edf-8c36-42b82ce49e2d for instance with vm_state building and task_state spawning. [ 983.685303] env[65385]: DEBUG nova.compute.manager [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-changed-14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 983.685463] env[65385]: DEBUG nova.compute.manager [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing instance network info cache due to event network-changed-14fd5163-1262-4edf-8c36-42b82ce49e2d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 983.685637] env[65385]: DEBUG oslo_concurrency.lockutils [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.758081] env[65385]: DEBUG nova.network.neutron [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 984.110604] env[65385]: DEBUG nova.scheduler.client.report [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 984.149679] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493267} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.150082] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 984.150342] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.150689] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-839cbd57-68cf-4ec3-9b24-484aa1aeab73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.158737] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 984.158737] env[65385]: value = "task-4454154" [ 984.158737] env[65385]: _type = "Task" [ 984.158737] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.168528] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.261384] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 984.261790] env[65385]: DEBUG nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Instance network_info: |[{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 984.262156] env[65385]: DEBUG oslo_concurrency.lockutils [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 984.262390] env[65385]: DEBUG nova.network.neutron [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing network info cache for port 14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 984.265027] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:d2:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14fd5163-1262-4edf-8c36-42b82ce49e2d', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.271321] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 984.272047] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.272278] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c01ae42-7f3b-4f72-bbcd-2d0e8a9cf27c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.290929] env[65385]: DEBUG nova.compute.manager [req-f6e16178-6586-4f94-8e67-6ea653e034ab req-16d928e4-01aa-4055-a577-adcc34d8e256 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received event network-vif-plugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 984.290929] env[65385]: DEBUG oslo_concurrency.lockutils [req-f6e16178-6586-4f94-8e67-6ea653e034ab req-16d928e4-01aa-4055-a577-adcc34d8e256 service nova] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.290929] env[65385]: DEBUG oslo_concurrency.lockutils [req-f6e16178-6586-4f94-8e67-6ea653e034ab req-16d928e4-01aa-4055-a577-adcc34d8e256 service nova] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.290929] env[65385]: DEBUG oslo_concurrency.lockutils [req-f6e16178-6586-4f94-8e67-6ea653e034ab req-16d928e4-01aa-4055-a577-adcc34d8e256 service nova] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.290929] env[65385]: DEBUG nova.compute.manager [req-f6e16178-6586-4f94-8e67-6ea653e034ab req-16d928e4-01aa-4055-a577-adcc34d8e256 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] No waiting events found dispatching network-vif-plugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 984.290929] env[65385]: WARNING nova.compute.manager [req-f6e16178-6586-4f94-8e67-6ea653e034ab req-16d928e4-01aa-4055-a577-adcc34d8e256 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received unexpected event network-vif-plugged-bbadc19c-78ee-4246-a53d-77f01c5ccea8 for instance with vm_state shelved_offloaded and task_state spawning. [ 984.298204] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.298204] env[65385]: value = "task-4454155" [ 984.298204] env[65385]: _type = "Task" [ 984.298204] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.308611] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454155, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.370913] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.370913] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 984.370913] env[65385]: DEBUG nova.network.neutron [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 984.620016] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.813s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.622472] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.537s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.622750] env[65385]: DEBUG nova.objects.instance [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'resources' on Instance uuid 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.640354] env[65385]: INFO nova.scheduler.client.report [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Deleted allocations for instance 55936da6-fcdf-4291-9b3a-23fa8b7a56ef [ 984.670782] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062003} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.671074] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.672048] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e56d35-612f-4f61-a40f-d8ba96d7e30a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.693280] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.693535] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86636ed9-b537-4873-a261-07189f780e47 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.716239] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 984.716239] env[65385]: value = "task-4454156" [ 984.716239] env[65385]: _type = "Task" [ 984.716239] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.725413] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454156, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.772789] env[65385]: WARNING neutronclient.v2_0.client [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 984.773604] env[65385]: WARNING openstack [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 984.774013] env[65385]: WARNING openstack [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 984.808910] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454155, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.876562] env[65385]: WARNING neutronclient.v2_0.client [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 984.877314] env[65385]: WARNING openstack [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 984.877719] env[65385]: WARNING openstack [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 984.887881] env[65385]: WARNING openstack [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 984.887881] env[65385]: WARNING openstack [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 984.953262] env[65385]: WARNING neutronclient.v2_0.client [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 984.953920] env[65385]: WARNING openstack [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 984.954290] env[65385]: WARNING openstack [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 984.986307] env[65385]: WARNING openstack [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 984.986713] env[65385]: WARNING openstack [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.048026] env[65385]: DEBUG nova.network.neutron [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updated VIF entry in instance network info cache for port 14fd5163-1262-4edf-8c36-42b82ce49e2d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 985.048026] env[65385]: DEBUG nova.network.neutron [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 985.056608] env[65385]: WARNING neutronclient.v2_0.client [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.057568] env[65385]: WARNING openstack [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.057743] env[65385]: WARNING openstack [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.129493] env[65385]: DEBUG nova.objects.instance [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'numa_topology' on Instance uuid 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.148359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69c69a26-779b-4745-86ec-bf81f298b40b tempest-InstanceActionsV221TestJSON-1222154335 tempest-InstanceActionsV221TestJSON-1222154335-project-member] Lock "55936da6-fcdf-4291-9b3a-23fa8b7a56ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.728s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 985.151481] env[65385]: DEBUG nova.network.neutron [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbadc19c-78", "ovs_interfaceid": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 985.226896] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454156, 'name': ReconfigVM_Task, 'duration_secs': 0.298852} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.227199] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6/7b6bab87-d2a3-4131-896f-3e4fb197ecd6.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.227986] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96972f90-d1e5-442b-a4d1-50086da6297c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.235087] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 985.235087] env[65385]: value = "task-4454157" [ 985.235087] env[65385]: _type = "Task" [ 985.235087] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.246797] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454157, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.310925] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454155, 'name': CreateVM_Task, 'duration_secs': 0.872316} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.311223] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.312034] env[65385]: WARNING neutronclient.v2_0.client [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.312639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.312881] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.313453] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 985.314025] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd70f525-e09b-4f97-a372-cd7b76ced9bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.323456] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 985.323456] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5289fcd2-c7ea-c795-691a-d7b764217809" [ 985.323456] env[65385]: _type = "Task" [ 985.323456] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.336338] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5289fcd2-c7ea-c795-691a-d7b764217809, 'name': SearchDatastore_Task, 'duration_secs': 0.010728} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.336680] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 985.336942] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.337225] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.337385] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.337566] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.337829] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a0903da-a7bf-465e-8d8f-c34543e3f969 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.347352] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.347537] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.348485] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28719b15-4233-4d9a-adf3-85e5ab0930b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.356075] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 985.356075] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525c1a16-a452-e32c-add7-9413d1237acd" [ 985.356075] env[65385]: _type = "Task" [ 985.356075] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.364561] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525c1a16-a452-e32c-add7-9413d1237acd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.551085] env[65385]: DEBUG oslo_concurrency.lockutils [req-25564350-204c-45f7-860b-e9352f0c3690 req-f0f90bae-03f4-46cd-9250-8d743b4c8b59 service nova] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 985.632235] env[65385]: DEBUG nova.objects.base [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Object Instance<048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7> lazy-loaded attributes: resources,numa_topology {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 985.653990] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 985.681906] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='4dd1180988c82146eefd9b029c88a06c',container_format='bare',created_at=2025-11-14T16:52:53Z,direct_url=,disk_format='vmdk',id=c4887e2f-7d54-4a01-8b97-eb6bbf0384e8,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-2144972100-shelved',owner='647f4e3157c34a8f9a64cb8ba8b72da6',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2025-11-14T16:53:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 985.681906] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 985.681906] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 985.681906] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 985.682161] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 985.682212] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 985.682396] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.682551] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 985.682735] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 985.682915] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 985.683077] env[65385]: DEBUG nova.virt.hardware [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 985.683909] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea69edc-79cf-469d-9ac0-51c09c12fc02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.695568] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb68235-e7d3-40f3-91ca-3abc3aa80fb8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.712984] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:ff:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4734e5e-2a76-4bda-8905-70c9bf9e007f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bbadc19c-78ee-4246-a53d-77f01c5ccea8', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.720619] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 985.723859] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.724146] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-627c111e-a65a-4ffa-932b-ffa1fdff2904 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.754018] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454157, 'name': Rename_Task, 'duration_secs': 0.146632} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.758628] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 985.758890] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.758890] env[65385]: value = "task-4454158" [ 985.758890] env[65385]: _type = "Task" [ 985.758890] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.759323] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb8df776-7ae2-4310-8c31-5f544dff2c99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.771777] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454158, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.777294] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 985.777294] env[65385]: value = "task-4454159" [ 985.777294] env[65385]: _type = "Task" [ 985.777294] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.788865] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454159, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.863583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0b87e6-d5bc-486e-884e-4cc7c92dbb76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.873878] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525c1a16-a452-e32c-add7-9413d1237acd, 'name': SearchDatastore_Task, 'duration_secs': 0.019972} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.877150] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdebdf56-d962-4041-8a30-12b14be60ae5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.880763] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74978d52-de53-4eef-a5c0-bdcbb41cad48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.891725] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 985.891725] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ba228e-9ea2-8dd1-7d5b-894608b64c5a" [ 985.891725] env[65385]: _type = "Task" [ 985.891725] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.921158] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6136ad-aff4-4547-831e-d9eeae95a3d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.930977] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ba228e-9ea2-8dd1-7d5b-894608b64c5a, 'name': SearchDatastore_Task, 'duration_secs': 0.011457} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.935919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 985.935919] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] a5fe2ad9-6dfe-485b-bf27-beb218d31467/a5fe2ad9-6dfe-485b-bf27-beb218d31467.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 985.935919] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1dd76bde-33af-4cf1-9595-ced1baddb556 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.937276] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89dabce7-6cbb-4f30-99f9-0846a8428242 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.953333] env[65385]: DEBUG nova.compute.provider_tree [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.957129] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 985.957129] env[65385]: value = "task-4454160" [ 985.957129] env[65385]: _type = "Task" [ 985.957129] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.968804] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454160, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.277882] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454158, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.287015] env[65385]: DEBUG oslo_vmware.api [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454159, 'name': PowerOnVM_Task, 'duration_secs': 0.500973} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.287264] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 986.287471] env[65385]: DEBUG nova.compute.manager [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 986.288883] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751ccfe3-c837-4ca6-aa75-881e19714659 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.313026] env[65385]: DEBUG nova.compute.manager [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received event network-changed-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 986.313181] env[65385]: DEBUG nova.compute.manager [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Refreshing instance network info cache due to event network-changed-bbadc19c-78ee-4246-a53d-77f01c5ccea8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 986.313936] env[65385]: DEBUG oslo_concurrency.lockutils [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Acquiring lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.313936] env[65385]: DEBUG oslo_concurrency.lockutils [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Acquired lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 986.313936] env[65385]: DEBUG nova.network.neutron [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Refreshing network info cache for port bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 986.462350] env[65385]: DEBUG nova.scheduler.client.report [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 986.479672] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454160, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524344} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.480092] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] a5fe2ad9-6dfe-485b-bf27-beb218d31467/a5fe2ad9-6dfe-485b-bf27-beb218d31467.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.480220] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.480785] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae303e2d-ad84-4136-b921-1632da8cd2dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.488789] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 986.488789] env[65385]: value = "task-4454161" [ 986.488789] env[65385]: _type = "Task" [ 986.488789] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.502062] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454161, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.773405] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454158, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.807944] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 986.816583] env[65385]: WARNING neutronclient.v2_0.client [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 986.817296] env[65385]: WARNING openstack [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.817641] env[65385]: WARNING openstack [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.948813] env[65385]: WARNING openstack [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.949207] env[65385]: WARNING openstack [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.970342] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.348s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 986.979782] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.062s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 986.980066] env[65385]: DEBUG nova.objects.instance [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lazy-loading 'resources' on Instance uuid f4af288e-b9c9-4273-8e65-e726ae039d03 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.015025] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454161, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068768} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.015025] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.015025] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b051efc-9009-4e50-a058-669d6644613f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.039918] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] a5fe2ad9-6dfe-485b-bf27-beb218d31467/a5fe2ad9-6dfe-485b-bf27-beb218d31467.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.040970] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-168e1ece-6736-44d9-91cd-2ed389687b17 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.062707] env[65385]: WARNING neutronclient.v2_0.client [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.063051] env[65385]: WARNING openstack [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.063460] env[65385]: WARNING openstack [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.079886] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 987.079886] env[65385]: value = "task-4454162" [ 987.079886] env[65385]: _type = "Task" [ 987.079886] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.091686] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454162, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.167388] env[65385]: DEBUG nova.network.neutron [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updated VIF entry in instance network info cache for port bbadc19c-78ee-4246-a53d-77f01c5ccea8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 987.168211] env[65385]: DEBUG nova.network.neutron [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbadc19c-78", "ovs_interfaceid": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 987.277012] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454158, 'name': CreateVM_Task, 'duration_secs': 1.403285} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.277294] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.278018] env[65385]: WARNING neutronclient.v2_0.client [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.278543] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.278763] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 987.279267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 987.279627] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83478ef2-3838-4c6d-a43c-fa312dc2025e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.286492] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 987.286492] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc50cd-8cf2-0aea-ffdf-73392e9a6a89" [ 987.286492] env[65385]: _type = "Task" [ 987.286492] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.300558] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bc50cd-8cf2-0aea-ffdf-73392e9a6a89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.486790] env[65385]: DEBUG oslo_concurrency.lockutils [None req-09b4a337-8018-42d1-949a-8d3831f5d47e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.707s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.489029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 11.014s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.489029] env[65385]: INFO nova.compute.manager [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Unshelving [ 987.593957] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454162, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.671649] env[65385]: DEBUG oslo_concurrency.lockutils [req-e866f2c4-99a3-4ff6-a219-33ad43945810 req-cd8ad586-f4c9-4ae4-80ec-345c00718bb3 service nova] Releasing lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 987.779139] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13869151-e75c-48b6-8e8f-91d0f03f09c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.787305] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff72b33-3357-41b8-9e80-9fb9156eee6f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.824949] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 987.825293] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Processing image c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.825539] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.825683] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 987.825862] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.826414] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c829681-c396-46bc-829a-6568e28c6914 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.829131] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89c99f5-8f80-4a94-a119-2c19ab8ab7cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.838696] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb16b1e-5331-4392-bb35-db9c47568db1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.855751] env[65385]: DEBUG nova.compute.provider_tree [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.859605] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.859605] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.859605] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43112f16-8ef8-48d4-8ee6-213bebecabce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.866611] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 987.866611] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9a164-bf5b-4856-de2e-76ce56c12e09" [ 987.866611] env[65385]: _type = "Task" [ 987.866611] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.875353] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9a164-bf5b-4856-de2e-76ce56c12e09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.936919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 987.936919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.936919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 987.937102] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.937256] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.939339] env[65385]: INFO nova.compute.manager [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Terminating instance [ 988.092066] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454162, 'name': ReconfigVM_Task, 'duration_secs': 0.54896} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.092283] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfigured VM instance instance-00000058 to attach disk [datastore2] a5fe2ad9-6dfe-485b-bf27-beb218d31467/a5fe2ad9-6dfe-485b-bf27-beb218d31467.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.092946] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0bd2e723-0171-4019-ac97-6271a3501aea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.100598] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 988.100598] env[65385]: value = "task-4454163" [ 988.100598] env[65385]: _type = "Task" [ 988.100598] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.111512] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454163, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.361332] env[65385]: DEBUG nova.scheduler.client.report [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 988.378881] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Preparing fetch location {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 988.379252] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Fetch image to [datastore1] OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477/OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477.vmdk {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 988.379464] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Downloading stream optimized image c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 to [datastore1] OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477/OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477.vmdk on the data store datastore1 as vApp {{(pid=65385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 988.379672] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Downloading image file data c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 to the ESX as VM named 'OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477' {{(pid=65385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 988.444852] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "refresh_cache-7b6bab87-d2a3-4131-896f-3e4fb197ecd6" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.445106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquired lock "refresh_cache-7b6bab87-d2a3-4131-896f-3e4fb197ecd6" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.445342] env[65385]: DEBUG nova.network.neutron [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 988.466960] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 988.466960] env[65385]: value = "resgroup-9" [ 988.466960] env[65385]: _type = "ResourcePool" [ 988.466960] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 988.467274] env[65385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-12b220d9-4294-4a8b-ba7c-e99f41421d42 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.492836] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lease: (returnval){ [ 988.492836] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5296d5be-a8b7-75e3-cb5c-0fcb121ebdd9" [ 988.492836] env[65385]: _type = "HttpNfcLease" [ 988.492836] env[65385]: } obtained for vApp import into resource pool (val){ [ 988.492836] env[65385]: value = "resgroup-9" [ 988.492836] env[65385]: _type = "ResourcePool" [ 988.492836] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 988.493152] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the lease: (returnval){ [ 988.493152] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5296d5be-a8b7-75e3-cb5c-0fcb121ebdd9" [ 988.493152] env[65385]: _type = "HttpNfcLease" [ 988.493152] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 988.500256] env[65385]: DEBUG nova.compute.utils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 988.506262] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 988.506262] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5296d5be-a8b7-75e3-cb5c-0fcb121ebdd9" [ 988.506262] env[65385]: _type = "HttpNfcLease" [ 988.506262] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 988.614462] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454163, 'name': Rename_Task, 'duration_secs': 0.160565} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.615097] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.615097] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec15df2c-da89-416e-843c-af657bda7279 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.625023] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 988.625023] env[65385]: value = "task-4454165" [ 988.625023] env[65385]: _type = "Task" [ 988.625023] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.633844] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454165, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.866701] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 988.869516] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.649s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.869817] env[65385]: DEBUG nova.objects.instance [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lazy-loading 'resources' on Instance uuid 3febf6af-699a-4fa1-b079-3790cd1095e5 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.893744] env[65385]: INFO nova.scheduler.client.report [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Deleted allocations for instance f4af288e-b9c9-4273-8e65-e726ae039d03 [ 988.947606] env[65385]: WARNING neutronclient.v2_0.client [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.948648] env[65385]: WARNING openstack [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.949207] env[65385]: WARNING openstack [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.976543] env[65385]: DEBUG nova.network.neutron [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 989.006009] env[65385]: INFO nova.virt.block_device [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Booting with volume fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2 at /dev/sdb [ 989.007810] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 989.007810] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5296d5be-a8b7-75e3-cb5c-0fcb121ebdd9" [ 989.007810] env[65385]: _type = "HttpNfcLease" [ 989.007810] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 989.040405] env[65385]: DEBUG nova.network.neutron [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 989.050410] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e3587c3-df27-4065-819a-634ad64aa675 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.064854] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c86bf5b-7a9a-41c5-8388-df39af9bc1df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.099106] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6516bb0a-e520-40ce-a986-6b30a0450940 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.109796] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196368d0-ccc7-425a-aa9a-4f5d11dc8e1c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.134504] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454165, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.143852] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7978662b-ce3d-4a8f-990b-17bae99d61b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.152568] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bfe8a8-0fdf-43db-9d1d-33b6f6ecbb94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.167884] env[65385]: DEBUG nova.virt.block_device [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating existing volume attachment record: 8486e3ce-0fa0-465e-b22d-70dc7d162e26 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 989.407026] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ea995a28-18f5-456f-a72f-e30a73010666 tempest-ImagesTestJSON-597039775 tempest-ImagesTestJSON-597039775-project-member] Lock "f4af288e-b9c9-4273-8e65-e726ae039d03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.070s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 989.503607] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 989.503607] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5296d5be-a8b7-75e3-cb5c-0fcb121ebdd9" [ 989.503607] env[65385]: _type = "HttpNfcLease" [ 989.503607] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 989.503975] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 989.503975] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5296d5be-a8b7-75e3-cb5c-0fcb121ebdd9" [ 989.503975] env[65385]: _type = "HttpNfcLease" [ 989.503975] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 989.505046] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866a294d-cf91-4d08-af2e-18538b0701d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.517783] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b88ac8-9182-cef8-735f-7f560a25f609/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 989.518074] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b88ac8-9182-cef8-735f-7f560a25f609/disk-0.vmdk. {{(pid=65385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 989.578343] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Releasing lock "refresh_cache-7b6bab87-d2a3-4131-896f-3e4fb197ecd6" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 989.578815] env[65385]: DEBUG nova.compute.manager [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 989.579027] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.583024] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd80f6d-5892-4ea5-89d0-31f465568018 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.597071] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-eee8b379-292b-4262-9440-8aa7cd45ec36 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.598798] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.600650] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-893e5d80-c605-4a60-a97d-ce6bb834fb0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.613105] env[65385]: DEBUG oslo_vmware.api [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 989.613105] env[65385]: value = "task-4454169" [ 989.613105] env[65385]: _type = "Task" [ 989.613105] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.623141] env[65385]: DEBUG oslo_vmware.api [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.637904] env[65385]: DEBUG oslo_vmware.api [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454165, 'name': PowerOnVM_Task, 'duration_secs': 0.81723} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.638111] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.638540] env[65385]: INFO nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Took 8.11 seconds to spawn the instance on the hypervisor. [ 989.638742] env[65385]: DEBUG nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 989.639605] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f1d9a8-d65f-4624-958e-de13d9502575 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.709446] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7272c24b-9baf-41a1-b851-9599919d6cc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.724041] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122f87b0-7413-4356-a808-22b16140d0d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.759825] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcd7e6e-3947-4787-8f21-dde7bcf15f76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.770552] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf70020-72d6-4a5c-a6e1-c0c675ba3379 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.787573] env[65385]: DEBUG nova.compute.provider_tree [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.125871] env[65385]: DEBUG oslo_vmware.api [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454169, 'name': PowerOffVM_Task, 'duration_secs': 0.139373} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.127760] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.127969] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.128281] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b69697b1-6b29-4d58-8e67-c29d7487098b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.161060] env[65385]: INFO nova.compute.manager [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Took 23.92 seconds to build instance. [ 990.172441] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.172669] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.172933] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Deleting the datastore file [datastore2] 7b6bab87-d2a3-4131-896f-3e4fb197ecd6 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.173257] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9533d9b-a156-4d5f-bc0f-12493b05f742 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.188406] env[65385]: DEBUG oslo_vmware.api [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for the task: (returnval){ [ 990.188406] env[65385]: value = "task-4454171" [ 990.188406] env[65385]: _type = "Task" [ 990.188406] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.204635] env[65385]: DEBUG oslo_vmware.api [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.296080] env[65385]: DEBUG nova.scheduler.client.report [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 990.673426] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d623d6f6-9235-436d-936b-f32faf6d9aa1 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.445s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.701531] env[65385]: DEBUG oslo_vmware.api [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Task: {'id': task-4454171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123242} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.701531] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.701531] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.701806] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.701806] env[65385]: INFO nova.compute.manager [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 990.701995] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 990.702233] env[65385]: DEBUG nova.compute.manager [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 990.702332] env[65385]: DEBUG nova.network.neutron [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 990.702582] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.703174] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.703430] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.805237] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.935s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.812330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.173s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.812330] env[65385]: DEBUG nova.objects.instance [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'resources' on Instance uuid c003dd80-d419-4a5a-8f86-dcac1a1e789c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.846620] env[65385]: INFO nova.scheduler.client.report [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Deleted allocations for instance 3febf6af-699a-4fa1-b079-3790cd1095e5 [ 990.859544] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Completed reading data from the image iterator. {{(pid=65385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 990.859764] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b88ac8-9182-cef8-735f-7f560a25f609/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 990.862353] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b4a01f-bb5c-4b87-bfea-4110f04e78b5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.874053] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b88ac8-9182-cef8-735f-7f560a25f609/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 990.874264] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b88ac8-9182-cef8-735f-7f560a25f609/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 990.874527] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-c980320c-fea8-4261-8576-ac98731c3aaa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.938296] env[65385]: DEBUG nova.network.neutron [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 990.938296] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.132068] env[65385]: DEBUG oslo_vmware.rw_handles [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b88ac8-9182-cef8-735f-7f560a25f609/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 991.132368] env[65385]: INFO nova.virt.vmwareapi.images [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Downloaded image file data c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 [ 991.133928] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfb19f0-1684-4198-a7cb-b6f31c41dc7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.160403] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b89ff5c4-d051-4631-8d20-90e2116dd8f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.188198] env[65385]: INFO nova.virt.vmwareapi.images [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] The imported VM was unregistered [ 991.195501] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Caching image {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 991.195801] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Creating directory with path [datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.196168] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38072cfc-29c1-49ba-b96a-a3fac42c0d8a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.213447] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Created directory with path [datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8 {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.213447] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477/OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477.vmdk to [datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8.vmdk. {{(pid=65385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 991.213677] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-26e317ed-ea02-4b74-8717-f816ab198da4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.225062] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 991.225062] env[65385]: value = "task-4454173" [ 991.225062] env[65385]: _type = "Task" [ 991.225062] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.231405] env[65385]: DEBUG nova.compute.manager [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-changed-14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 991.231821] env[65385]: DEBUG nova.compute.manager [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing instance network info cache due to event network-changed-14fd5163-1262-4edf-8c36-42b82ce49e2d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 991.232347] env[65385]: DEBUG oslo_concurrency.lockutils [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.232347] env[65385]: DEBUG oslo_concurrency.lockutils [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.232485] env[65385]: DEBUG nova.network.neutron [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing network info cache for port 14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 991.241852] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454173, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.360935] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba9d8e32-3f54-4090-a657-50cf8e2072dd tempest-ServersTestJSON-1079632421 tempest-ServersTestJSON-1079632421-project-member] Lock "3febf6af-699a-4fa1-b079-3790cd1095e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.489s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 991.441716] env[65385]: DEBUG nova.network.neutron [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 991.622997] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28f84ad-5c59-4379-8058-d95f56dc9c23 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.636231] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53795f0-16fc-4a40-bfb0-c69aaa1f0218 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.683660] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b88d199-868e-4c6f-8363-36326f89dbb2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.696081] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810b5a5e-b56d-48f2-8f12-2e994a4ae4f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.717025] env[65385]: DEBUG nova.compute.provider_tree [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.736276] env[65385]: WARNING neutronclient.v2_0.client [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.736985] env[65385]: WARNING openstack [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.738055] env[65385]: WARNING openstack [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.752226] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454173, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.946721] env[65385]: INFO nova.compute.manager [-] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Took 1.24 seconds to deallocate network for instance. [ 992.163494] env[65385]: WARNING openstack [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 992.164829] env[65385]: WARNING openstack [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 992.223242] env[65385]: DEBUG nova.scheduler.client.report [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 992.243110] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454173, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.368839] env[65385]: WARNING neutronclient.v2_0.client [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 992.369717] env[65385]: WARNING openstack [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 992.372618] env[65385]: WARNING openstack [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 992.456440] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 992.713314] env[65385]: DEBUG nova.network.neutron [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updated VIF entry in instance network info cache for port 14fd5163-1262-4edf-8c36-42b82ce49e2d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 992.714064] env[65385]: DEBUG nova.network.neutron [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 992.734288] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.737285] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.898s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 992.737555] env[65385]: DEBUG nova.objects.instance [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 992.746479] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454173, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.773248] env[65385]: INFO nova.scheduler.client.report [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted allocations for instance c003dd80-d419-4a5a-8f86-dcac1a1e789c [ 993.219032] env[65385]: DEBUG oslo_concurrency.lockutils [req-89fb2b5b-c26d-4eca-885f-1f7dd3dbeb95 req-b39e6108-e566-4ec3-88dc-504b66988a69 service nova] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 993.246170] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454173, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.285719] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67df473c-4a13-4feb-bc42-09d853ca0acc tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "c003dd80-d419-4a5a-8f86-dcac1a1e789c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.590s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.565967] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 993.566512] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.745647] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454173, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.754236] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e443cc4-5d29-441e-94ad-269ca5084eb2 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.755998] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.418s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.759624] env[65385]: INFO nova.compute.claims [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.069613] env[65385]: DEBUG nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 994.243436] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454173, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.60752} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.244466] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477/OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477.vmdk to [datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8.vmdk. [ 994.244466] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Cleaning up location [datastore1] OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 994.244466] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_47b0b415-62c8-4fcb-af50-1fdded1c9477 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.244702] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-486cdfe7-d3cd-4e03-b63e-79f4cccfd9e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.252621] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 994.252621] env[65385]: value = "task-4454175" [ 994.252621] env[65385]: _type = "Task" [ 994.252621] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.272247] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454175, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.606711] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.768324] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454175, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038354} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.769910] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.769910] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.769910] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8.vmdk to [datastore1] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.773043] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea376a4b-cd42-4253-bdc9-a826af7927e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.784456] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 994.784456] env[65385]: value = "task-4454176" [ 994.784456] env[65385]: _type = "Task" [ 994.784456] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.801914] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454176, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.823470] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.008340] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c81f140-3dbe-4765-8163-3ed9f090f506 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.018038] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d57b19-0207-4029-ae4a-221a201e4142 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.055051] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017f3a9b-9d46-4ada-8065-af5d1cd00b65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.066980] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7432373f-ba66-4ecf-a414-0145efeea615 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.091095] env[65385]: DEBUG nova.compute.provider_tree [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.265178] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "8a974e99-a7b1-438e-aace-9ae82352495c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.265551] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.301151] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454176, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.594299] env[65385]: DEBUG nova.scheduler.client.report [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.767527] env[65385]: DEBUG nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 995.800312] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454176, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.100074] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.344s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.100660] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 996.105594] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.740s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.108135] env[65385]: INFO nova.compute.claims [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.292488] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.301052] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454176, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.608088] env[65385]: DEBUG nova.compute.utils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 996.610758] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 996.610758] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 996.611207] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.611870] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.612787] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.613530] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.681833] env[65385]: DEBUG nova.policy [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd122bc776500423ab3e717a0f4d5229e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abc72fbdb7e0400b8ce8b8ef0f946fb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 996.755476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "537393de-ea48-4d34-8e04-95331bbc099d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.755714] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "537393de-ea48-4d34-8e04-95331bbc099d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.801689] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454176, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.022427] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Successfully created port: d974d069-fef1-4319-a1f9-6accdbd536b5 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 997.125791] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 997.259099] env[65385]: DEBUG nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 997.303834] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454176, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.435621} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.309481] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8/c4887e2f-7d54-4a01-8b97-eb6bbf0384e8.vmdk to [datastore1] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.312510] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58913f45-2b53-4ba7-af83-b2642cde1563 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.342562] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.346233] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3377464-2df4-4665-9b18-cea448c3c2f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.368653] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 997.368653] env[65385]: value = "task-4454177" [ 997.368653] env[65385]: _type = "Task" [ 997.368653] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.383888] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454177, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.445437] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb4e64d-ff66-4694-a36b-b3d05ce59dfa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.454496] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79971993-8987-434a-b5c5-433363eecaa6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.489202] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3925262c-ec08-4982-b836-a6ebf2fbb39f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.498520] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd86ed6-e50d-4bfa-827e-fc2c92ee26eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.513693] env[65385]: DEBUG nova.compute.provider_tree [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.659020] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.659290] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.659646] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.659646] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.659785] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.660331] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.660331] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.660718] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 997.660718] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 997.784747] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.879079] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454177, 'name': ReconfigVM_Task, 'duration_secs': 0.283439} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.879373] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 0c345794-036c-45d1-985e-cf5c116bd873/0c345794-036c-45d1-985e-cf5c116bd873.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.880012] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6702c02-4a4e-4f9d-9127-c569b8ef20cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.887993] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 997.887993] env[65385]: value = "task-4454178" [ 997.887993] env[65385]: _type = "Task" [ 997.887993] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.896835] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454178, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.016962] env[65385]: DEBUG nova.scheduler.client.report [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 998.137376] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 998.163653] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.168182] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 998.168451] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 998.168724] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 998.168964] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 998.169160] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 998.169284] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 998.169466] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.169620] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 998.169782] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 998.169923] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 998.170100] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 998.170965] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3ac677-ea8d-4f2a-bae8-89b40adf6058 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.179608] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9960be9-913e-493c-9679-726aeca6e8d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.398683] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454178, 'name': Rename_Task, 'duration_secs': 0.160466} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.398991] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.399953] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c0ed7a6-2d11-4146-ac06-ea5b0b234381 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.407579] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 998.407579] env[65385]: value = "task-4454179" [ 998.407579] env[65385]: _type = "Task" [ 998.407579] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.418585] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454179, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.522213] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.528475] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 998.530197] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.464s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.531070] env[65385]: DEBUG nova.objects.instance [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lazy-loading 'resources' on Instance uuid 167e1b46-af52-4a7a-9964-edabc8155287 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.539134] env[65385]: DEBUG nova.compute.manager [req-00c41f23-0ef3-4569-b383-3aa12b5b8584 req-dc8dbe02-ed23-4dcd-983c-2fd20876691f service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Received event network-vif-plugged-d974d069-fef1-4319-a1f9-6accdbd536b5 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 998.539134] env[65385]: DEBUG oslo_concurrency.lockutils [req-00c41f23-0ef3-4569-b383-3aa12b5b8584 req-dc8dbe02-ed23-4dcd-983c-2fd20876691f service nova] Acquiring lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.539308] env[65385]: DEBUG oslo_concurrency.lockutils [req-00c41f23-0ef3-4569-b383-3aa12b5b8584 req-dc8dbe02-ed23-4dcd-983c-2fd20876691f service nova] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.539525] env[65385]: DEBUG oslo_concurrency.lockutils [req-00c41f23-0ef3-4569-b383-3aa12b5b8584 req-dc8dbe02-ed23-4dcd-983c-2fd20876691f service nova] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.539713] env[65385]: DEBUG nova.compute.manager [req-00c41f23-0ef3-4569-b383-3aa12b5b8584 req-dc8dbe02-ed23-4dcd-983c-2fd20876691f service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] No waiting events found dispatching network-vif-plugged-d974d069-fef1-4319-a1f9-6accdbd536b5 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 998.539873] env[65385]: WARNING nova.compute.manager [req-00c41f23-0ef3-4569-b383-3aa12b5b8584 req-dc8dbe02-ed23-4dcd-983c-2fd20876691f service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Received unexpected event network-vif-plugged-d974d069-fef1-4319-a1f9-6accdbd536b5 for instance with vm_state building and task_state spawning. [ 998.635621] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Successfully updated port: d974d069-fef1-4319-a1f9-6accdbd536b5 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 998.918445] env[65385]: DEBUG oslo_vmware.api [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454179, 'name': PowerOnVM_Task, 'duration_secs': 0.497174} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.918734] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.037728] env[65385]: DEBUG nova.compute.utils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 999.039349] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 999.039468] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 999.040228] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.040228] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.040632] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.040983] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.049834] env[65385]: DEBUG nova.compute.manager [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 999.051243] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72cb7a1-cee9-43a7-9e65-41be7aa87f10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.117194] env[65385]: DEBUG nova.policy [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd122bc776500423ab3e717a0f4d5229e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abc72fbdb7e0400b8ce8b8ef0f946fb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 999.138600] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "refresh_cache-cb99e6ec-fefc-4f44-86a0-9d886e9c722d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.138849] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "refresh_cache-cb99e6ec-fefc-4f44-86a0-9d886e9c722d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 999.138964] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 999.277367] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a3d0df-db9d-4693-a8db-a027d3c2d7db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.289483] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5a7990-4fe3-4b60-b0c1-8fd3cc5a6c6b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.324885] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b804cb5c-94cc-47e8-a610-0f7db00b1c4f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.333859] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a495ca3d-b7b7-4118-82b6-0025c8a33f32 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.349045] env[65385]: DEBUG nova.compute.provider_tree [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.450332] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Successfully created port: c958eb48-c18c-4b17-99dd-ea1d002a2b4c {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 999.550403] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 999.572794] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55c29d2f-99ce-4aab-87f4-af90472de122 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 28.880s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.645022] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.645022] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.684061] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 999.727770] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.728188] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.804880] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.805678] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.806044] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.852058] env[65385]: DEBUG nova.scheduler.client.report [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 999.929377] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Updating instance_info_cache with network_info: [{"id": "d974d069-fef1-4319-a1f9-6accdbd536b5", "address": "fa:16:3e:5e:1f:58", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd974d069-fe", "ovs_interfaceid": "d974d069-fef1-4319-a1f9-6accdbd536b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1000.359065] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.360673] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.118s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.361250] env[65385]: DEBUG nova.objects.instance [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'resources' on Instance uuid 26a2d464-0aab-463f-8265-948a4f0ff188 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.383034] env[65385]: INFO nova.scheduler.client.report [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Deleted allocations for instance 167e1b46-af52-4a7a-9964-edabc8155287 [ 1000.431917] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "refresh_cache-cb99e6ec-fefc-4f44-86a0-9d886e9c722d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1000.432555] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Instance network_info: |[{"id": "d974d069-fef1-4319-a1f9-6accdbd536b5", "address": "fa:16:3e:5e:1f:58", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd974d069-fe", "ovs_interfaceid": "d974d069-fef1-4319-a1f9-6accdbd536b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1000.433689] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:1f:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd974d069-fef1-4319-a1f9-6accdbd536b5', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.441444] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1000.443333] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1000.446226] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d15871ab-0122-4af8-b02b-5f9400c8c9b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.468949] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4cf85b-a2f9-4a23-99dc-e196c26b3fd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.473556] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.473556] env[65385]: value = "task-4454180" [ 1000.473556] env[65385]: _type = "Task" [ 1000.473556] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.479417] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-126a8c1b-db17-4f67-9bc6-4db84b975178 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Suspending the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1000.480157] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ac682ef5-26e5-43c3-a7cf-1edfb82c9528 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.485483] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454180, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.491095] env[65385]: DEBUG oslo_vmware.api [None req-126a8c1b-db17-4f67-9bc6-4db84b975178 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 1000.491095] env[65385]: value = "task-4454181" [ 1000.491095] env[65385]: _type = "Task" [ 1000.491095] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.501246] env[65385]: DEBUG oslo_vmware.api [None req-126a8c1b-db17-4f67-9bc6-4db84b975178 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454181, 'name': SuspendVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.561079] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1000.591639] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1000.591882] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1000.592119] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1000.592229] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1000.592452] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1000.592642] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1000.592907] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.593194] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1000.593397] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1000.593565] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1000.593796] env[65385]: DEBUG nova.virt.hardware [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1000.594792] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045668d0-2fd4-4b15-b8a9-139b9a6675d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.604035] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f583a963-5884-45c1-b638-4d459c946d50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.659767] env[65385]: DEBUG nova.compute.manager [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Received event network-changed-d974d069-fef1-4319-a1f9-6accdbd536b5 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1000.660097] env[65385]: DEBUG nova.compute.manager [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Refreshing instance network info cache due to event network-changed-d974d069-fef1-4319-a1f9-6accdbd536b5. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1000.660468] env[65385]: DEBUG oslo_concurrency.lockutils [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Acquiring lock "refresh_cache-cb99e6ec-fefc-4f44-86a0-9d886e9c722d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.660552] env[65385]: DEBUG oslo_concurrency.lockutils [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Acquired lock "refresh_cache-cb99e6ec-fefc-4f44-86a0-9d886e9c722d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1000.660721] env[65385]: DEBUG nova.network.neutron [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Refreshing network info cache for port d974d069-fef1-4319-a1f9-6accdbd536b5 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1000.894110] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e87d16d-3b5e-45ef-96bf-01a49c2a995c tempest-AttachInterfacesV270Test-147385612 tempest-AttachInterfacesV270Test-147385612-project-member] Lock "167e1b46-af52-4a7a-9964-edabc8155287" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.284s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.941215] env[65385]: DEBUG nova.compute.manager [req-be524396-71c3-4f18-a282-6478e2dce5e7 req-ba170140-9be4-4eb5-9a45-700fbf186b5b service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Received event network-vif-plugged-c958eb48-c18c-4b17-99dd-ea1d002a2b4c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1000.941464] env[65385]: DEBUG oslo_concurrency.lockutils [req-be524396-71c3-4f18-a282-6478e2dce5e7 req-ba170140-9be4-4eb5-9a45-700fbf186b5b service nova] Acquiring lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.942077] env[65385]: DEBUG oslo_concurrency.lockutils [req-be524396-71c3-4f18-a282-6478e2dce5e7 req-ba170140-9be4-4eb5-9a45-700fbf186b5b service nova] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.942280] env[65385]: DEBUG oslo_concurrency.lockutils [req-be524396-71c3-4f18-a282-6478e2dce5e7 req-ba170140-9be4-4eb5-9a45-700fbf186b5b service nova] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.942510] env[65385]: DEBUG nova.compute.manager [req-be524396-71c3-4f18-a282-6478e2dce5e7 req-ba170140-9be4-4eb5-9a45-700fbf186b5b service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] No waiting events found dispatching network-vif-plugged-c958eb48-c18c-4b17-99dd-ea1d002a2b4c {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1000.942698] env[65385]: WARNING nova.compute.manager [req-be524396-71c3-4f18-a282-6478e2dce5e7 req-ba170140-9be4-4eb5-9a45-700fbf186b5b service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Received unexpected event network-vif-plugged-c958eb48-c18c-4b17-99dd-ea1d002a2b4c for instance with vm_state building and task_state spawning. [ 1000.991330] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454180, 'name': CreateVM_Task, 'duration_secs': 0.462556} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.991628] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1000.996229] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1000.996229] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.996539] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1000.996713] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1000.997324] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4469b3d4-2163-41d0-a300-47fd3e61fe90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.006641] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1001.006641] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f7fe6c-8353-7046-a9c3-da909af964a6" [ 1001.006641] env[65385]: _type = "Task" [ 1001.006641] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.011150] env[65385]: DEBUG oslo_vmware.api [None req-126a8c1b-db17-4f67-9bc6-4db84b975178 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454181, 'name': SuspendVM_Task} progress is 58%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.026695] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f7fe6c-8353-7046-a9c3-da909af964a6, 'name': SearchDatastore_Task, 'duration_secs': 0.012133} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.027065] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1001.029721] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.030015] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.031020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1001.031020] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.033584] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f960031-d246-4c0e-8aa5-6b776e62e9d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.044377] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.044684] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1001.048030] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19691ef2-de09-4a67-a331-b5e5d4390932 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.049835] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Successfully updated port: c958eb48-c18c-4b17-99dd-ea1d002a2b4c {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1001.060695] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1001.060695] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dd4cd3-ce15-ee07-e69f-744b272454a2" [ 1001.060695] env[65385]: _type = "Task" [ 1001.060695] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.074019] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dd4cd3-ce15-ee07-e69f-744b272454a2, 'name': SearchDatastore_Task, 'duration_secs': 0.009805} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.074019] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c416746-a44e-48ca-981a-b6fe64f3808c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.084226] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1001.084226] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52107985-9949-c26d-ead7-7244e2c02872" [ 1001.084226] env[65385]: _type = "Task" [ 1001.084226] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.095863] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52107985-9949-c26d-ead7-7244e2c02872, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.103752] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aaab0ce-a5b6-45ff-a64f-8af27abe57b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.111908] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5face670-9a47-4ed5-ab03-0632c7551737 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.146788] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d7abbb-0450-4d00-a7f6-8f6e2632625b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.155836] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de596ff-bcdd-453a-ab2b-d5b4109d6f97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.171267] env[65385]: WARNING neutronclient.v2_0.client [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.172930] env[65385]: WARNING openstack [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.172930] env[65385]: WARNING openstack [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.181444] env[65385]: DEBUG nova.compute.provider_tree [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.309181] env[65385]: WARNING openstack [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.309734] env[65385]: WARNING openstack [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.370259] env[65385]: WARNING neutronclient.v2_0.client [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.370913] env[65385]: WARNING openstack [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.371367] env[65385]: WARNING openstack [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.449906] env[65385]: DEBUG nova.network.neutron [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Updated VIF entry in instance network info cache for port d974d069-fef1-4319-a1f9-6accdbd536b5. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1001.450880] env[65385]: DEBUG nova.network.neutron [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Updating instance_info_cache with network_info: [{"id": "d974d069-fef1-4319-a1f9-6accdbd536b5", "address": "fa:16:3e:5e:1f:58", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd974d069-fe", "ovs_interfaceid": "d974d069-fef1-4319-a1f9-6accdbd536b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1001.503170] env[65385]: DEBUG oslo_vmware.api [None req-126a8c1b-db17-4f67-9bc6-4db84b975178 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454181, 'name': SuspendVM_Task, 'duration_secs': 0.695837} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.503547] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-126a8c1b-db17-4f67-9bc6-4db84b975178 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Suspended the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1001.503733] env[65385]: DEBUG nova.compute.manager [None req-126a8c1b-db17-4f67-9bc6-4db84b975178 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1001.504621] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fa77c4-cfb2-4dd2-a3a6-100328e9e03e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.555670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "refresh_cache-6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.555670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "refresh_cache-6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1001.555670] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1001.594710] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52107985-9949-c26d-ead7-7244e2c02872, 'name': SearchDatastore_Task, 'duration_secs': 0.010687} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.594963] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1001.595240] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] cb99e6ec-fefc-4f44-86a0-9d886e9c722d/cb99e6ec-fefc-4f44-86a0-9d886e9c722d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1001.595500] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93fb69fc-4597-4fe3-9005-95438b6a615c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.603997] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1001.603997] env[65385]: value = "task-4454182" [ 1001.603997] env[65385]: _type = "Task" [ 1001.603997] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.612673] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454182, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.688316] env[65385]: DEBUG nova.scheduler.client.report [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1001.952615] env[65385]: DEBUG oslo_concurrency.lockutils [req-2b11ed51-822f-43e3-8ff8-34adb7052636 req-82afc5fb-9b21-46e8-92f4-4061cd644946 service nova] Releasing lock "refresh_cache-cb99e6ec-fefc-4f44-86a0-9d886e9c722d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1002.060454] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.060866] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.100450] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1002.114665] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454182, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.121614] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.122012] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.184638] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.185381] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.185877] env[65385]: WARNING openstack [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.195132] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.199398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.390s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.199398] env[65385]: DEBUG nova.objects.instance [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1002.235048] env[65385]: INFO nova.scheduler.client.report [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted allocations for instance 26a2d464-0aab-463f-8265-948a4f0ff188 [ 1002.352193] env[65385]: DEBUG nova.network.neutron [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Updating instance_info_cache with network_info: [{"id": "c958eb48-c18c-4b17-99dd-ea1d002a2b4c", "address": "fa:16:3e:0d:89:62", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc958eb48-c1", "ovs_interfaceid": "c958eb48-c18c-4b17-99dd-ea1d002a2b4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1002.408524] env[65385]: INFO nova.compute.manager [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Resuming [ 1002.408524] env[65385]: DEBUG nova.objects.instance [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lazy-loading 'flavor' on Instance uuid 0c345794-036c-45d1-985e-cf5c116bd873 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.616739] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454182, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738396} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.617196] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] cb99e6ec-fefc-4f44-86a0-9d886e9c722d/cb99e6ec-fefc-4f44-86a0-9d886e9c722d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1002.617511] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.617511] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0dad9467-6dfd-4997-bf0e-c276ac658bd5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.625226] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1002.625226] env[65385]: value = "task-4454183" [ 1002.625226] env[65385]: _type = "Task" [ 1002.625226] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.634826] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.748250] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbe9b39d-e2ea-41c5-ac1f-3ece852c34de tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "26a2d464-0aab-463f-8265-948a4f0ff188" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.151s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.857428] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "refresh_cache-6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1002.857428] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Instance network_info: |[{"id": "c958eb48-c18c-4b17-99dd-ea1d002a2b4c", "address": "fa:16:3e:0d:89:62", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc958eb48-c1", "ovs_interfaceid": "c958eb48-c18c-4b17-99dd-ea1d002a2b4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1002.857428] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:89:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c958eb48-c18c-4b17-99dd-ea1d002a2b4c', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.868787] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1002.868839] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.870354] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c73c5a0a-6c28-4bda-821f-c6b37bc4d171 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.892494] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.892494] env[65385]: value = "task-4454184" [ 1002.892494] env[65385]: _type = "Task" [ 1002.892494] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.905460] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454184, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.973930] env[65385]: DEBUG nova.compute.manager [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Received event network-changed-c958eb48-c18c-4b17-99dd-ea1d002a2b4c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1002.973930] env[65385]: DEBUG nova.compute.manager [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Refreshing instance network info cache due to event network-changed-c958eb48-c18c-4b17-99dd-ea1d002a2b4c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1002.974211] env[65385]: DEBUG oslo_concurrency.lockutils [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Acquiring lock "refresh_cache-6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.974459] env[65385]: DEBUG oslo_concurrency.lockutils [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Acquired lock "refresh_cache-6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1002.974734] env[65385]: DEBUG nova.network.neutron [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Refreshing network info cache for port c958eb48-c18c-4b17-99dd-ea1d002a2b4c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1003.136887] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068764} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.137194] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.137998] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d833c031-cca2-4910-839b-70260595805b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.162901] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] cb99e6ec-fefc-4f44-86a0-9d886e9c722d/cb99e6ec-fefc-4f44-86a0-9d886e9c722d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.163174] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aacab105-58b7-40f3-982f-59951ae50b28 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.183896] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1003.183896] env[65385]: value = "task-4454185" [ 1003.183896] env[65385]: _type = "Task" [ 1003.183896] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.193367] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.208108] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed35a526-22fc-471b-9cb5-e640a45de072 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.209492] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.754s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.210114] env[65385]: DEBUG nova.objects.instance [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lazy-loading 'resources' on Instance uuid 7b6bab87-d2a3-4131-896f-3e4fb197ecd6 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.406103] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454184, 'name': CreateVM_Task, 'duration_secs': 0.396935} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.406103] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.406103] env[65385]: WARNING neutronclient.v2_0.client [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.406667] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.406667] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1003.406861] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1003.407094] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cf8a7f4-a5e2-42f6-92a2-3971a9cd0d44 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.412599] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1003.412599] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52330a2c-ffe6-258b-8b94-998d0fc058b6" [ 1003.412599] env[65385]: _type = "Task" [ 1003.412599] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.424886] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52330a2c-ffe6-258b-8b94-998d0fc058b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.477017] env[65385]: WARNING neutronclient.v2_0.client [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.477856] env[65385]: WARNING openstack [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.478281] env[65385]: WARNING openstack [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.696990] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454185, 'name': ReconfigVM_Task} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.698081] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Reconfigured VM instance instance-00000059 to attach disk [datastore1] cb99e6ec-fefc-4f44-86a0-9d886e9c722d/cb99e6ec-fefc-4f44-86a0-9d886e9c722d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.698478] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79ae8859-a551-42e4-abaa-dc4ae6dda15b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.707949] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1003.707949] env[65385]: value = "task-4454186" [ 1003.707949] env[65385]: _type = "Task" [ 1003.707949] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.718397] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454186, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.924704] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.924886] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquired lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1003.926681] env[65385]: DEBUG nova.network.neutron [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1003.940409] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52330a2c-ffe6-258b-8b94-998d0fc058b6, 'name': SearchDatastore_Task, 'duration_secs': 0.010097} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.941402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1003.941638] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.941866] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.942019] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1003.943066] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.943066] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee515681-62a7-48e1-9ead-15f2e0d8fe6a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.956393] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.956589] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.957441] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d449a93-78ea-4068-9b47-f99d6df059d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.962437] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c97e7e1-b33d-403f-892b-566d4a4e7f64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.974901] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1003.974901] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ed8b51-eff9-b946-46e5-bdaa7061c40f" [ 1003.974901] env[65385]: _type = "Task" [ 1003.974901] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.981819] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7a1e8d-c56a-4d9f-88ff-00e8cba453eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.992117] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ed8b51-eff9-b946-46e5-bdaa7061c40f, 'name': SearchDatastore_Task, 'duration_secs': 0.010335} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.023851] env[65385]: WARNING openstack [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.023851] env[65385]: WARNING openstack [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.029524] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a7ed97f-6646-4902-9c4f-016a272f919c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.032691] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6705a4fa-8044-4133-ad23-3391a0a04555 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.043671] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87d3c63-a2c0-4d87-9e7e-c6d3b0993264 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.048143] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1004.048143] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c58341-06a6-5302-9659-81ff6908bd03" [ 1004.048143] env[65385]: _type = "Task" [ 1004.048143] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.060736] env[65385]: DEBUG nova.compute.provider_tree [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.068842] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c58341-06a6-5302-9659-81ff6908bd03, 'name': SearchDatastore_Task, 'duration_secs': 0.013796} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.069757] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1004.070142] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5/6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.070447] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c773f12d-3a4a-4134-a010-73c7deafb649 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.079360] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1004.079360] env[65385]: value = "task-4454187" [ 1004.079360] env[65385]: _type = "Task" [ 1004.079360] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.090058] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.223621] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454186, 'name': Rename_Task, 'duration_secs': 0.149871} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.223931] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1004.224256] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e32ff167-6cdf-4a06-a045-1eee044d4605 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.232917] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1004.232917] env[65385]: value = "task-4454188" [ 1004.232917] env[65385]: _type = "Task" [ 1004.232917] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.242933] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.377314] env[65385]: WARNING neutronclient.v2_0.client [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1004.377999] env[65385]: WARNING openstack [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.378407] env[65385]: WARNING openstack [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.429773] env[65385]: WARNING neutronclient.v2_0.client [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1004.430653] env[65385]: WARNING openstack [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.432144] env[65385]: WARNING openstack [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.564753] env[65385]: DEBUG nova.scheduler.client.report [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1004.592325] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454187, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.743903] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454188, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.884513] env[65385]: DEBUG nova.network.neutron [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Updated VIF entry in instance network info cache for port c958eb48-c18c-4b17-99dd-ea1d002a2b4c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1004.885217] env[65385]: DEBUG nova.network.neutron [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Updating instance_info_cache with network_info: [{"id": "c958eb48-c18c-4b17-99dd-ea1d002a2b4c", "address": "fa:16:3e:0d:89:62", "network": {"id": "13be2a6b-52e0-4b1b-9df6-b86381dc05b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2133723135-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abc72fbdb7e0400b8ce8b8ef0f946fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc958eb48-c1", "ovs_interfaceid": "c958eb48-c18c-4b17-99dd-ea1d002a2b4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1005.074661] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.077711] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.471s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.079371] env[65385]: INFO nova.compute.claims [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.104109] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454187, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518137} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.105191] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5/6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.105191] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.105191] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ceb2b0e-7b4b-4e43-beb8-4864d158d893 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.118370] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1005.118370] env[65385]: value = "task-4454189" [ 1005.118370] env[65385]: _type = "Task" [ 1005.118370] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.119608] env[65385]: INFO nova.scheduler.client.report [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Deleted allocations for instance 7b6bab87-d2a3-4131-896f-3e4fb197ecd6 [ 1005.135949] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454189, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.140389] env[65385]: WARNING openstack [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.140875] env[65385]: WARNING openstack [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.252622] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454188, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.391656] env[65385]: DEBUG oslo_concurrency.lockutils [req-fe3589e5-14ee-43db-8b73-a4f449fd6ef2 req-4eaa042f-fbae-4204-bf0d-52a76fb755cd service nova] Releasing lock "refresh_cache-6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1005.632146] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454189, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142969} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.634782] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.635381] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a4bc9d5-0a16-47d1-b0ca-d70e9ac37b77 tempest-ServerShowV257Test-655156488 tempest-ServerShowV257Test-655156488-project-member] Lock "7b6bab87-d2a3-4131-896f-3e4fb197ecd6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.699s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.637527] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe22e7e-2254-4f45-b3a2-45fd56fb1f2f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.667789] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5/6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.668170] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a4a3398-7951-486b-942a-c86428856ac2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.691154] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1005.691154] env[65385]: value = "task-4454190" [ 1005.691154] env[65385]: _type = "Task" [ 1005.691154] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.703514] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454190, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.745590] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454188, 'name': PowerOnVM_Task, 'duration_secs': 1.13178} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.746079] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1005.746163] env[65385]: INFO nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Took 7.61 seconds to spawn the instance on the hypervisor. [ 1005.746339] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1005.747217] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fe19ff-84b5-4187-a609-c85176018426 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.780178] env[65385]: WARNING neutronclient.v2_0.client [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.781228] env[65385]: WARNING openstack [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.781228] env[65385]: WARNING openstack [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.075652] env[65385]: DEBUG nova.network.neutron [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [{"id": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "address": "fa:16:3e:53:ff:f9", "network": {"id": "4f89a392-dc77-4f3f-88ac-0152367fcaba", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2098584550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "647f4e3157c34a8f9a64cb8ba8b72da6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbadc19c-78", "ovs_interfaceid": "bbadc19c-78ee-4246-a53d-77f01c5ccea8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1006.208590] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454190, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.269147] env[65385]: INFO nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Took 25.95 seconds to build instance. [ 1006.450315] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f55fe2-e7d7-45f4-b70f-893b66b75dec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.463752] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9f8a4d-fded-4b43-b605-af885139807e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.518497] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32348a43-2087-46f5-a0a9-f79cf1029332 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.528809] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313fd69b-bdc5-43f7-8b71-699727731441 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.547730] env[65385]: DEBUG nova.compute.provider_tree [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.579732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Releasing lock "refresh_cache-0c345794-036c-45d1-985e-cf5c116bd873" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1006.580742] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54149ab7-9ac0-4dfc-9d67-4728e3f4b6ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.592892] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Resuming the VM {{(pid=65385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1006.593009] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6952f71b-c530-4c97-b5f3-7783cc705c46 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.603025] env[65385]: DEBUG oslo_vmware.api [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 1006.603025] env[65385]: value = "task-4454191" [ 1006.603025] env[65385]: _type = "Task" [ 1006.603025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.616606] env[65385]: DEBUG oslo_vmware.api [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.705142] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454190, 'name': ReconfigVM_Task, 'duration_secs': 0.828624} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.705573] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5/6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.706543] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1db4f228-c4d2-46a2-9b8a-116247250728 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.715534] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1006.715534] env[65385]: value = "task-4454192" [ 1006.715534] env[65385]: _type = "Task" [ 1006.715534] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.728636] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454192, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.772219] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.462s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.051670] env[65385]: DEBUG nova.scheduler.client.report [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1007.118983] env[65385]: DEBUG oslo_vmware.api [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454191, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.211178] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.211573] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.233843] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454192, 'name': Rename_Task, 'duration_secs': 0.155122} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.234886] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.235174] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73760dd5-f975-4bef-a458-416bc686e96d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.243064] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1007.243064] env[65385]: value = "task-4454193" [ 1007.243064] env[65385]: _type = "Task" [ 1007.243064] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.253800] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.557211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.557768] env[65385]: DEBUG nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1007.561520] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.738s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.562433] env[65385]: DEBUG nova.objects.instance [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'pci_requests' on Instance uuid 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.614810] env[65385]: DEBUG oslo_vmware.api [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454191, 'name': PowerOnVM_Task, 'duration_secs': 0.586995} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.615092] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Resumed the VM {{(pid=65385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1007.615276] env[65385]: DEBUG nova.compute.manager [None req-63266d13-28f3-418f-a62a-530777180136 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1007.616379] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbbcafb-4486-4719-b702-a343b1682606 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.705880] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "992920c3-af4f-489e-b21f-9d52c24399de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.706159] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "992920c3-af4f-489e-b21f-9d52c24399de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.715147] env[65385]: DEBUG nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1007.753520] env[65385]: DEBUG oslo_vmware.api [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454193, 'name': PowerOnVM_Task, 'duration_secs': 0.498364} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.753798] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.754012] env[65385]: INFO nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Took 7.19 seconds to spawn the instance on the hypervisor. [ 1007.754196] env[65385]: DEBUG nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1007.755383] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841a2393-5a49-4706-9b52-5eb730600e55 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.066385] env[65385]: DEBUG nova.compute.utils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1008.069674] env[65385]: DEBUG nova.objects.instance [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'numa_topology' on Instance uuid 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.070939] env[65385]: DEBUG nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1008.071241] env[65385]: DEBUG nova.network.neutron [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1008.071743] env[65385]: WARNING neutronclient.v2_0.client [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.072121] env[65385]: WARNING neutronclient.v2_0.client [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.072776] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.073184] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.081557] env[65385]: INFO nova.compute.claims [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.120842] env[65385]: DEBUG nova.policy [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b450c500f84f2fa682d70698c398b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d45ff3446a84ef2a1945f7a66b747e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1008.208308] env[65385]: DEBUG nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1008.247527] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.273087] env[65385]: INFO nova.compute.manager [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Took 27.92 seconds to build instance. [ 1008.519303] env[65385]: DEBUG nova.network.neutron [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Successfully created port: c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1008.572118] env[65385]: DEBUG nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1008.630504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.631127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.631607] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.632029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.632366] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.636034] env[65385]: INFO nova.compute.manager [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Terminating instance [ 1008.718154] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.735656] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.777954] env[65385]: DEBUG oslo_concurrency.lockutils [None req-af5ccb21-bf87-4310-a924-fad0309ef1ca tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.435s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.778291] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.060s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.778513] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.778774] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.778950] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.783168] env[65385]: INFO nova.compute.manager [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Terminating instance [ 1009.139943] env[65385]: DEBUG nova.compute.manager [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1009.140242] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.141028] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c4674e-2abc-4099-b7c4-f3696bef806a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.149567] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.152307] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8380f889-73c3-4912-87a4-7c3600cb090f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.161013] env[65385]: DEBUG oslo_vmware.api [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1009.161013] env[65385]: value = "task-4454194" [ 1009.161013] env[65385]: _type = "Task" [ 1009.161013] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.173618] env[65385]: DEBUG oslo_vmware.api [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.287070] env[65385]: DEBUG nova.compute.manager [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1009.287321] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.288180] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a663362f-a27e-4911-856e-0118a537ee25 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.291629] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f6ad1f-e4fc-4d65-95bc-bc6fbee98ad7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.301964] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b48502b-d7d2-449d-86dd-4f8b7f99e335 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.305221] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.305474] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b2cac83-e4c8-4253-9430-a06382625fca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.339215] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9dae66-81ae-4864-baf0-fea0ee141640 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.342436] env[65385]: DEBUG oslo_vmware.api [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1009.342436] env[65385]: value = "task-4454195" [ 1009.342436] env[65385]: _type = "Task" [ 1009.342436] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.353761] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5aec77-ffd4-4a27-a5d6-f963c3fa25f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.364527] env[65385]: DEBUG oslo_vmware.api [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.380036] env[65385]: DEBUG nova.compute.provider_tree [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.583124] env[65385]: DEBUG nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1009.610632] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1009.610884] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1009.611050] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1009.611268] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1009.611427] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1009.611525] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1009.611721] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.611875] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1009.612057] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1009.612261] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1009.612388] env[65385]: DEBUG nova.virt.hardware [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1009.613304] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9362fef8-822f-41e0-91ed-9345e295af50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.622214] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08cf6fc-c35e-4adf-a527-3d20bbd6d3b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.671301] env[65385]: DEBUG oslo_vmware.api [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454194, 'name': PowerOffVM_Task, 'duration_secs': 0.228374} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.671621] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.671792] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.672066] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52b497f1-82a7-4631-849a-c1001b7df0d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.740079] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.740079] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.740324] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleting the datastore file [datastore1] cb99e6ec-fefc-4f44-86a0-9d886e9c722d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.740551] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5ac304d-e51a-437d-be0b-b6e89d5ebb3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.748497] env[65385]: DEBUG oslo_vmware.api [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1009.748497] env[65385]: value = "task-4454197" [ 1009.748497] env[65385]: _type = "Task" [ 1009.748497] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.757411] env[65385]: DEBUG oslo_vmware.api [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.855095] env[65385]: DEBUG oslo_vmware.api [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454195, 'name': PowerOffVM_Task, 'duration_secs': 0.225043} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.855095] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.855095] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.855095] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fba07cca-75e9-4430-96d1-fc9ed24a6eb3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.883518] env[65385]: DEBUG nova.scheduler.client.report [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1010.061257] env[65385]: DEBUG nova.compute.manager [req-6ab887ec-2894-4e12-b712-4d2a307e7c2f req-5a43cb01-a0a7-405a-aac7-f2a403d9c8bd service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Received event network-vif-plugged-c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1010.061439] env[65385]: DEBUG oslo_concurrency.lockutils [req-6ab887ec-2894-4e12-b712-4d2a307e7c2f req-5a43cb01-a0a7-405a-aac7-f2a403d9c8bd service nova] Acquiring lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.061664] env[65385]: DEBUG oslo_concurrency.lockutils [req-6ab887ec-2894-4e12-b712-4d2a307e7c2f req-5a43cb01-a0a7-405a-aac7-f2a403d9c8bd service nova] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.061858] env[65385]: DEBUG oslo_concurrency.lockutils [req-6ab887ec-2894-4e12-b712-4d2a307e7c2f req-5a43cb01-a0a7-405a-aac7-f2a403d9c8bd service nova] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.062063] env[65385]: DEBUG nova.compute.manager [req-6ab887ec-2894-4e12-b712-4d2a307e7c2f req-5a43cb01-a0a7-405a-aac7-f2a403d9c8bd service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] No waiting events found dispatching network-vif-plugged-c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1010.063886] env[65385]: WARNING nova.compute.manager [req-6ab887ec-2894-4e12-b712-4d2a307e7c2f req-5a43cb01-a0a7-405a-aac7-f2a403d9c8bd service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Received unexpected event network-vif-plugged-c11bab1f-7eb9-434c-8436-c2319951b01f for instance with vm_state building and task_state spawning. [ 1010.136025] env[65385]: DEBUG nova.network.neutron [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Successfully updated port: c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1010.259979] env[65385]: DEBUG oslo_vmware.api [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265745} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.260391] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.260391] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.260519] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.260762] env[65385]: INFO nova.compute.manager [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1010.260945] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1010.261158] env[65385]: DEBUG nova.compute.manager [-] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1010.261259] env[65385]: DEBUG nova.network.neutron [-] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1010.261539] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1010.262092] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1010.262357] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1010.301043] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1010.389302] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.828s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.389773] env[65385]: WARNING neutronclient.v2_0.client [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1010.392866] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.100s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.394662] env[65385]: INFO nova.compute.claims [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.422538] env[65385]: INFO nova.network.neutron [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating port f7b3d6ae-eb2d-43c1-8c44-ca85849be867 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1010.564430] env[65385]: DEBUG nova.compute.manager [req-8d711f4b-84ab-4f23-8f22-e3db1c421e9b req-68a79cd3-37b7-4e44-b806-829be9d6917d service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Received event network-vif-deleted-d974d069-fef1-4319-a1f9-6accdbd536b5 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1010.564490] env[65385]: INFO nova.compute.manager [req-8d711f4b-84ab-4f23-8f22-e3db1c421e9b req-68a79cd3-37b7-4e44-b806-829be9d6917d service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Neutron deleted interface d974d069-fef1-4319-a1f9-6accdbd536b5; detaching it from the instance and deleting it from the info cache [ 1010.564644] env[65385]: DEBUG nova.network.neutron [req-8d711f4b-84ab-4f23-8f22-e3db1c421e9b req-68a79cd3-37b7-4e44-b806-829be9d6917d service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1010.589173] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "370b6aa8-2337-48e1-8102-611ce4494a55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.589426] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "370b6aa8-2337-48e1-8102-611ce4494a55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.589629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "370b6aa8-2337-48e1-8102-611ce4494a55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.589805] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "370b6aa8-2337-48e1-8102-611ce4494a55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.589964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "370b6aa8-2337-48e1-8102-611ce4494a55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.591851] env[65385]: INFO nova.compute.manager [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Terminating instance [ 1010.640519] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.640519] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1010.640682] env[65385]: DEBUG nova.network.neutron [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1011.034036] env[65385]: DEBUG nova.network.neutron [-] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1011.067385] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6bf56728-632e-4078-ba1f-90b1c7fef71b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.078440] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e842cb2-6664-4043-9e58-aa970254de72 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.096031] env[65385]: DEBUG nova.compute.manager [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1011.096266] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1011.097193] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376d5e41-7352-4d31-94ad-53fa7a7bf5ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.109522] env[65385]: DEBUG nova.compute.manager [req-8d711f4b-84ab-4f23-8f22-e3db1c421e9b req-68a79cd3-37b7-4e44-b806-829be9d6917d service nova] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Detach interface failed, port_id=d974d069-fef1-4319-a1f9-6accdbd536b5, reason: Instance cb99e6ec-fefc-4f44-86a0-9d886e9c722d could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1011.115375] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.115658] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbc81636-b45d-4c91-a329-d32d5122ae99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.123653] env[65385]: DEBUG oslo_vmware.api [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1011.123653] env[65385]: value = "task-4454199" [ 1011.123653] env[65385]: _type = "Task" [ 1011.123653] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.131203] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1011.131454] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1011.131627] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleting the datastore file [datastore1] 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.135170] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b5fd551-8ac6-4964-a387-9d3465d366bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.137693] env[65385]: DEBUG oslo_vmware.api [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.144841] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.146172] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.152719] env[65385]: DEBUG oslo_vmware.api [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for the task: (returnval){ [ 1011.152719] env[65385]: value = "task-4454200" [ 1011.152719] env[65385]: _type = "Task" [ 1011.152719] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.162379] env[65385]: DEBUG oslo_vmware.api [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.185802] env[65385]: DEBUG nova.network.neutron [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1011.208791] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.209384] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.278014] env[65385]: WARNING neutronclient.v2_0.client [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.278752] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.279473] env[65385]: WARNING openstack [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.420031] env[65385]: DEBUG nova.network.neutron [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updating instance_info_cache with network_info: [{"id": "c11bab1f-7eb9-434c-8436-c2319951b01f", "address": "fa:16:3e:ea:99:f4", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc11bab1f-7e", "ovs_interfaceid": "c11bab1f-7eb9-434c-8436-c2319951b01f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1011.540233] env[65385]: INFO nova.compute.manager [-] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Took 1.28 seconds to deallocate network for instance. [ 1011.598518] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e08135-e4b9-44fc-ab64-66d8eeb2a690 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.608853] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d3ce8b-fb21-40ec-895c-b0603fff4215 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.646255] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e660ca-7df3-4985-a89c-00ffe4213452 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.657083] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92937f76-3115-4c39-8a72-4835ffd00365 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.663469] env[65385]: DEBUG oslo_vmware.api [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454199, 'name': PowerOffVM_Task, 'duration_secs': 0.21253} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.664086] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.664263] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1011.665449] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-206ab889-c3cc-4043-b09d-c7a0e3c2d9a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.678107] env[65385]: DEBUG oslo_vmware.api [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Task: {'id': task-4454200, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156415} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.678570] env[65385]: DEBUG nova.compute.provider_tree [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.679971] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.680186] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.680361] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.680537] env[65385]: INFO nova.compute.manager [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Took 2.39 seconds to destroy the instance on the hypervisor. [ 1011.680774] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1011.681203] env[65385]: DEBUG nova.compute.manager [-] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1011.681318] env[65385]: DEBUG nova.network.neutron [-] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1011.681537] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.682058] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.682345] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.739761] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.744330] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1011.744592] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1011.744778] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleting the datastore file [datastore1] 370b6aa8-2337-48e1-8102-611ce4494a55 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.745063] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17f24af0-e55e-40b3-9cbd-a32906778c7d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.752798] env[65385]: DEBUG oslo_vmware.api [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1011.752798] env[65385]: value = "task-4454202" [ 1011.752798] env[65385]: _type = "Task" [ 1011.752798] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.762484] env[65385]: DEBUG oslo_vmware.api [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.922477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1011.923154] env[65385]: DEBUG nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Instance network_info: |[{"id": "c11bab1f-7eb9-434c-8436-c2319951b01f", "address": "fa:16:3e:ea:99:f4", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc11bab1f-7e", "ovs_interfaceid": "c11bab1f-7eb9-434c-8436-c2319951b01f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1011.923512] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:99:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c11bab1f-7eb9-434c-8436-c2319951b01f', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.932115] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating folder: Project (8d45ff3446a84ef2a1945f7a66b747e8). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1011.932505] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a08aedc4-7f2f-4e65-ae6e-70f6a9d84065 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.945658] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created folder: Project (8d45ff3446a84ef2a1945f7a66b747e8) in parent group-v870881. [ 1011.947237] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating folder: Instances. Parent ref: group-v871140. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1011.947237] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6cacf7ef-d24f-44d1-ad5c-1930cff8e204 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.959930] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created folder: Instances in parent group-v871140. [ 1011.960339] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1011.960895] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1011.960895] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49953429-e546-44ca-91c1-a89b0d1a74ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.982309] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.982309] env[65385]: value = "task-4454205" [ 1011.982309] env[65385]: _type = "Task" [ 1011.982309] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.993683] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454205, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.048151] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1012.083485] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.083771] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1012.083865] env[65385]: DEBUG nova.network.neutron [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1012.185131] env[65385]: DEBUG nova.scheduler.client.report [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.193088] env[65385]: DEBUG nova.compute.manager [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Received event network-changed-c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1012.193275] env[65385]: DEBUG nova.compute.manager [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Refreshing instance network info cache due to event network-changed-c11bab1f-7eb9-434c-8436-c2319951b01f. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1012.193673] env[65385]: DEBUG oslo_concurrency.lockutils [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Acquiring lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.193842] env[65385]: DEBUG oslo_concurrency.lockutils [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Acquired lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1012.193950] env[65385]: DEBUG nova.network.neutron [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Refreshing network info cache for port c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1012.264159] env[65385]: DEBUG oslo_vmware.api [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219001} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.264415] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.264616] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1012.264795] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1012.264999] env[65385]: INFO nova.compute.manager [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1012.265271] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1012.265476] env[65385]: DEBUG nova.compute.manager [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1012.265575] env[65385]: DEBUG nova.network.neutron [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1012.265829] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.266404] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.266700] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.311768] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.491985] env[65385]: DEBUG nova.network.neutron [-] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1012.507538] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454205, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.587262] env[65385]: WARNING neutronclient.v2_0.client [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.587857] env[65385]: WARNING openstack [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.588349] env[65385]: WARNING openstack [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.601559] env[65385]: DEBUG nova.compute.manager [req-4cf8e8cc-d871-4862-86e8-80075754136f req-5b51ff2b-e425-497b-aa17-8c5b40693fe9 service nova] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Received event network-vif-deleted-c958eb48-c18c-4b17-99dd-ea1d002a2b4c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1012.692028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1012.692504] env[65385]: DEBUG nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1012.697983] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.913s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.699490] env[65385]: INFO nova.compute.claims [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.702320] env[65385]: WARNING neutronclient.v2_0.client [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.702973] env[65385]: WARNING openstack [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.703344] env[65385]: WARNING openstack [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.832478] env[65385]: WARNING openstack [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.833161] env[65385]: WARNING openstack [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.905285] env[65385]: WARNING openstack [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.906495] env[65385]: WARNING openstack [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.923833] env[65385]: WARNING neutronclient.v2_0.client [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.924554] env[65385]: WARNING openstack [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.924921] env[65385]: WARNING openstack [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.968194] env[65385]: WARNING neutronclient.v2_0.client [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.969535] env[65385]: WARNING openstack [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.969898] env[65385]: WARNING openstack [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.994035] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454205, 'name': CreateVM_Task, 'duration_secs': 0.64618} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.994035] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1012.994211] env[65385]: WARNING neutronclient.v2_0.client [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.994596] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.994772] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1012.995104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1012.995388] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbffcf6f-5e62-4ab5-832b-445d7dd827dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.001797] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1013.001797] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c632e-51ab-c0e3-efbc-ecb43a6cdc14" [ 1013.001797] env[65385]: _type = "Task" [ 1013.001797] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.009561] env[65385]: INFO nova.compute.manager [-] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Took 1.33 seconds to deallocate network for instance. [ 1013.020981] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c632e-51ab-c0e3-efbc-ecb43a6cdc14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.028331] env[65385]: DEBUG nova.network.neutron [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1013.074408] env[65385]: DEBUG nova.network.neutron [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updated VIF entry in instance network info cache for port c11bab1f-7eb9-434c-8436-c2319951b01f. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1013.074408] env[65385]: DEBUG nova.network.neutron [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updating instance_info_cache with network_info: [{"id": "c11bab1f-7eb9-434c-8436-c2319951b01f", "address": "fa:16:3e:ea:99:f4", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc11bab1f-7e", "ovs_interfaceid": "c11bab1f-7eb9-434c-8436-c2319951b01f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1013.218563] env[65385]: DEBUG nova.compute.utils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1013.220182] env[65385]: DEBUG nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1013.220787] env[65385]: DEBUG nova.network.neutron [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1013.221105] env[65385]: WARNING neutronclient.v2_0.client [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1013.221612] env[65385]: WARNING neutronclient.v2_0.client [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1013.222518] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.223068] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.236080] env[65385]: DEBUG nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1013.275761] env[65385]: DEBUG nova.policy [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d5103a5769a4296973351194f49fed3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '754b1facaaa14501b2204c98e1d7a5a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1013.392449] env[65385]: DEBUG nova.network.neutron [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1013.515039] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c632e-51ab-c0e3-efbc-ecb43a6cdc14, 'name': SearchDatastore_Task, 'duration_secs': 0.01748} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.515039] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1013.515039] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.515305] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.515305] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.515767] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.515767] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-944e32eb-2a9c-4d42-a90e-3a2a8dcb5e9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.527112] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.527247] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.528239] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36b051f2-397d-45c7-8c5d-16a01a7f04bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.535586] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.536087] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1013.539025] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1013.539025] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bf7aac-285d-c163-87bf-039aa905b0ca" [ 1013.539025] env[65385]: _type = "Task" [ 1013.539025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.553952] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52bf7aac-285d-c163-87bf-039aa905b0ca, 'name': SearchDatastore_Task, 'duration_secs': 0.014158} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.555095] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2089be43-b833-46fd-8199-c7d93370bfe2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.562381] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1013.562381] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b14a0d-1889-e362-9b5c-c057e749e729" [ 1013.562381] env[65385]: _type = "Task" [ 1013.562381] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.573959] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b14a0d-1889-e362-9b5c-c057e749e729, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='01173e1ea95c241f4ac14c12bed8f26d',container_format='bare',created_at=2025-11-14T16:53:11Z,direct_url=,disk_format='vmdk',id=1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1083390871-shelved',owner='b27c092a0fd345ea953956469c21f8bb',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2025-11-14T16:53:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1013.579925] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1013.580457] env[65385]: DEBUG nova.virt.hardware [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1013.580500] env[65385]: DEBUG oslo_concurrency.lockutils [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Releasing lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1013.580870] env[65385]: DEBUG nova.compute.manager [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-vif-plugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1013.580945] env[65385]: DEBUG oslo_concurrency.lockutils [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.581100] env[65385]: DEBUG oslo_concurrency.lockutils [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.581263] env[65385]: DEBUG oslo_concurrency.lockutils [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.581420] env[65385]: DEBUG nova.compute.manager [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] No waiting events found dispatching network-vif-plugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1013.581580] env[65385]: WARNING nova.compute.manager [req-ef679983-ac2f-4e84-a4d9-2387a49f0082 req-c6c50b0d-54ff-4402-abc4-0c0e42fa08d7 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received unexpected event network-vif-plugged-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 for instance with vm_state shelved_offloaded and task_state spawning. [ 1013.582465] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655acddd-ec35-449c-b6eb-664e78ea9199 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.592815] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f971d712-a9ab-4cb3-83b4-82c04cd7e2cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.610527] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:13:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4a8fd90-153b-494f-b76a-299eb05c03f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7b3d6ae-eb2d-43c1-8c44-ca85849be867', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.620414] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1013.620752] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.621331] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a256e6a8-109e-4583-8dc9-2788eb18811a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.638193] env[65385]: DEBUG nova.network.neutron [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Successfully created port: f26f849c-4075-4ee1-b174-88ce8e66b69c {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1013.643359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.643724] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.643952] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "0c345794-036c-45d1-985e-cf5c116bd873-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.644231] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.644495] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.647725] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.647725] env[65385]: value = "task-4454206" [ 1013.647725] env[65385]: _type = "Task" [ 1013.647725] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.648257] env[65385]: INFO nova.compute.manager [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Terminating instance [ 1013.660802] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454206, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.895747] env[65385]: INFO nova.compute.manager [-] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Took 1.63 seconds to deallocate network for instance. [ 1013.911763] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4021c016-be2d-43b9-98fc-8d83c72e7415 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.920758] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e71308c-2987-448d-b257-43ad5c43956a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.956350] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091801e9-75f9-4b27-b011-97fcff49f5db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.965728] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4edf15-1354-4dff-b645-e48240a3d6c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.982335] env[65385]: DEBUG nova.compute.provider_tree [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.073595] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b14a0d-1889-e362-9b5c-c057e749e729, 'name': SearchDatastore_Task, 'duration_secs': 0.012974} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.073899] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.074174] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b4bca2c8-fc8f-471a-899b-aac648fee9fe/b4bca2c8-fc8f-471a-899b-aac648fee9fe.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.074774] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1017d64f-21be-46e9-81af-a8e0c9ed3674 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.082863] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1014.082863] env[65385]: value = "task-4454207" [ 1014.082863] env[65385]: _type = "Task" [ 1014.082863] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.093971] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454207, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.155602] env[65385]: DEBUG nova.compute.manager [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1014.155940] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.160631] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f20bb05-15a1-4b8c-8ee4-ee444331dd3b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.163114] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454206, 'name': CreateVM_Task, 'duration_secs': 0.414621} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.163300] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.164297] env[65385]: WARNING neutronclient.v2_0.client [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.164636] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.164799] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.165161] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1014.165429] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-425fe7e3-d002-438f-a92b-5be541bdc880 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.169795] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.170530] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ad91140-6899-4864-8ccd-20837ee22f3f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.174916] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1014.174916] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52147122-b76c-3d03-26cd-062c120b27b8" [ 1014.174916] env[65385]: _type = "Task" [ 1014.174916] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.179915] env[65385]: DEBUG oslo_vmware.api [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 1014.179915] env[65385]: value = "task-4454208" [ 1014.179915] env[65385]: _type = "Task" [ 1014.179915] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.187928] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52147122-b76c-3d03-26cd-062c120b27b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.191235] env[65385]: DEBUG oslo_vmware.api [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454208, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.235699] env[65385]: DEBUG nova.compute.manager [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1014.235884] env[65385]: DEBUG nova.compute.manager [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing instance network info cache due to event network-changed-f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1014.236130] env[65385]: DEBUG oslo_concurrency.lockutils [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Acquiring lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.236319] env[65385]: DEBUG oslo_concurrency.lockutils [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Acquired lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.236499] env[65385]: DEBUG nova.network.neutron [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Refreshing network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1014.249478] env[65385]: DEBUG nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1014.282575] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1014.282858] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1014.283009] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1014.283333] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1014.283528] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1014.283713] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1014.283939] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.284136] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1014.284310] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1014.284485] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1014.284656] env[65385]: DEBUG nova.virt.hardware [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1014.285886] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c17916-37d0-4a7a-ac37-798e13309c99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.294816] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c59d570-26c6-4500-abfd-4015ef912e8a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.407856] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.485788] env[65385]: DEBUG nova.scheduler.client.report [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1014.597657] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454207, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.696208] env[65385]: DEBUG oslo_vmware.api [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454208, 'name': PowerOffVM_Task, 'duration_secs': 0.236015} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.697431] env[65385]: DEBUG nova.compute.manager [req-71d66d66-c6a8-4988-98b5-1a7e06e8913e req-ecd89034-08ca-41cc-90a4-ebd549340d20 service nova] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Received event network-vif-deleted-722cc380-0268-4488-ad6c-462f3af7b938 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1014.697707] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.697915] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.698275] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.698404] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Processing image 1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.698627] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.698760] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.698928] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.699518] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ddc92f14-336f-40d1-98b2-bf47d39995e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.701255] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60b1c773-bd36-4d57-9a94-3b04243e4b91 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.720604] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.720862] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.721621] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-605f7b6f-b6d0-4cb7-b864-bcdb47889391 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.728043] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1014.728043] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e9f225-edd5-714f-183e-2bc9a0e609dc" [ 1014.728043] env[65385]: _type = "Task" [ 1014.728043] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.736868] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e9f225-edd5-714f-183e-2bc9a0e609dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.739564] env[65385]: WARNING neutronclient.v2_0.client [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.740219] env[65385]: WARNING openstack [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.740559] env[65385]: WARNING openstack [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.804955] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.805269] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.805452] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleting the datastore file [datastore1] 0c345794-036c-45d1-985e-cf5c116bd873 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.805759] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b64edf68-4dc3-445f-b26b-56f028806d51 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.813423] env[65385]: DEBUG oslo_vmware.api [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for the task: (returnval){ [ 1014.813423] env[65385]: value = "task-4454210" [ 1014.813423] env[65385]: _type = "Task" [ 1014.813423] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.822440] env[65385]: DEBUG oslo_vmware.api [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454210, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.869027] env[65385]: WARNING openstack [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.869027] env[65385]: WARNING openstack [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.931630] env[65385]: WARNING neutronclient.v2_0.client [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.932344] env[65385]: WARNING openstack [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.932738] env[65385]: WARNING openstack [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.989864] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.990413] env[65385]: DEBUG nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1014.993020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.829s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.993198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.993357] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1014.993692] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.746s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.995261] env[65385]: INFO nova.compute.claims [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.999718] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae77aa2-4121-41c2-98fa-fa312ff47072 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.009629] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f231a982-7e9c-4380-a71b-464c5e803c21 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.028597] env[65385]: DEBUG nova.network.neutron [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updated VIF entry in instance network info cache for port f7b3d6ae-eb2d-43c1-8c44-ca85849be867. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1015.030970] env[65385]: DEBUG nova.network.neutron [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [{"id": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "address": "fa:16:3e:cf:13:60", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7b3d6ae-eb", "ovs_interfaceid": "f7b3d6ae-eb2d-43c1-8c44-ca85849be867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.030970] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfe1a09-1a50-4fdf-b482-d1da172a271e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.045275] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73e9175-1c6f-44fe-87b5-bd84aae7c24e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.079020] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179850MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1015.079412] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.096205] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454207, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620976} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.096205] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b4bca2c8-fc8f-471a-899b-aac648fee9fe/b4bca2c8-fc8f-471a-899b-aac648fee9fe.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.096205] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.096205] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-068e8c38-325a-45be-b13e-1992233bc259 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.105604] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1015.105604] env[65385]: value = "task-4454211" [ 1015.105604] env[65385]: _type = "Task" [ 1015.105604] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.116130] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454211, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.188826] env[65385]: DEBUG nova.network.neutron [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Successfully updated port: f26f849c-4075-4ee1-b174-88ce8e66b69c {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1015.240034] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Preparing fetch location {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1015.240174] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Fetch image to [datastore1] OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82/OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82.vmdk {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1015.240300] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Downloading stream optimized image 1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa to [datastore1] OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82/OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82.vmdk on the data store datastore1 as vApp {{(pid=65385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1015.240475] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Downloading image file data 1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa to the ESX as VM named 'OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82' {{(pid=65385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1015.326494] env[65385]: DEBUG oslo_vmware.api [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Task: {'id': task-4454210, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150869} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.327627] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1015.327627] env[65385]: value = "resgroup-9" [ 1015.327627] env[65385]: _type = "ResourcePool" [ 1015.327627] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1015.327922] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.328140] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1015.328315] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1015.328488] env[65385]: INFO nova.compute.manager [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1015.328726] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1015.328924] env[65385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-fcad7ba0-97e7-4f59-8923-12b143596f54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.344159] env[65385]: DEBUG nova.compute.manager [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1015.344327] env[65385]: DEBUG nova.network.neutron [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1015.344521] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.345103] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.345365] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.358874] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease: (returnval){ [ 1015.358874] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523a94b8-2035-1589-b2b9-eb176bb1e25d" [ 1015.358874] env[65385]: _type = "HttpNfcLease" [ 1015.358874] env[65385]: } obtained for vApp import into resource pool (val){ [ 1015.358874] env[65385]: value = "resgroup-9" [ 1015.358874] env[65385]: _type = "ResourcePool" [ 1015.358874] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1015.359238] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the lease: (returnval){ [ 1015.359238] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523a94b8-2035-1589-b2b9-eb176bb1e25d" [ 1015.359238] env[65385]: _type = "HttpNfcLease" [ 1015.359238] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1015.366771] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1015.366771] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523a94b8-2035-1589-b2b9-eb176bb1e25d" [ 1015.366771] env[65385]: _type = "HttpNfcLease" [ 1015.366771] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1015.385913] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.501177] env[65385]: DEBUG nova.compute.utils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1015.502798] env[65385]: DEBUG nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1015.503023] env[65385]: DEBUG nova.network.neutron [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1015.503368] env[65385]: WARNING neutronclient.v2_0.client [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.503700] env[65385]: WARNING neutronclient.v2_0.client [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.504356] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.504750] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.535811] env[65385]: DEBUG oslo_concurrency.lockutils [req-ab5eb5ab-18c1-44b1-a11a-8a4ac4f7f7dc req-7767bed4-bb9b-4e2b-84e5-695ae488eaaf service nova] Releasing lock "refresh_cache-048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.569357] env[65385]: DEBUG nova.policy [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c1a9d4194964403a8d11abfad4c65a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94cc5bc221c0455ea760f5022db6bdc4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1015.618980] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454211, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.193709} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.619237] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.620217] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e933ea4f-c8f5-4d0b-b785-0342b7dcce22 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.649962] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] b4bca2c8-fc8f-471a-899b-aac648fee9fe/b4bca2c8-fc8f-471a-899b-aac648fee9fe.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.650364] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-305b96ab-e08a-456a-806e-2632644a3733 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.674049] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1015.674049] env[65385]: value = "task-4454213" [ 1015.674049] env[65385]: _type = "Task" [ 1015.674049] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.683998] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454213, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.692389] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.692389] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1015.692389] env[65385]: DEBUG nova.network.neutron [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1015.868695] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1015.868695] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523a94b8-2035-1589-b2b9-eb176bb1e25d" [ 1015.868695] env[65385]: _type = "HttpNfcLease" [ 1015.868695] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1015.899407] env[65385]: DEBUG nova.network.neutron [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Successfully created port: a7e0f7ab-d411-40cb-ae92-d58b922f22fe {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1016.012823] env[65385]: DEBUG nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1016.106026] env[65385]: DEBUG nova.network.neutron [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1016.183651] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454213, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.196690] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.197101] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.221112] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196c7cac-eede-45ea-b968-1d23657ccc07 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.229311] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4ead0d-c361-4936-ae7f-abfce2267d1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.234463] env[65385]: DEBUG nova.network.neutron [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1016.263600] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.264119] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.271107] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b2c8d9-db4b-4009-88db-84c8c33b2cea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.279908] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe191fbf-6901-46a9-a663-c1140d63d80f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.295989] env[65385]: DEBUG nova.compute.provider_tree [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.334943] env[65385]: WARNING neutronclient.v2_0.client [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.335668] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.336149] env[65385]: WARNING openstack [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.368833] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.368833] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523a94b8-2035-1589-b2b9-eb176bb1e25d" [ 1016.368833] env[65385]: _type = "HttpNfcLease" [ 1016.368833] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1016.369204] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1016.369204] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523a94b8-2035-1589-b2b9-eb176bb1e25d" [ 1016.369204] env[65385]: _type = "HttpNfcLease" [ 1016.369204] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1016.369944] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699b3cb6-cecf-487a-88c7-4d3f0b3303e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.378355] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525c588e-814b-4dbb-a07f-e0ca1506bd88/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1016.378498] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525c588e-814b-4dbb-a07f-e0ca1506bd88/disk-0.vmdk. {{(pid=65385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1016.453584] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-98f214ab-96e1-4c15-9ce5-4142e4fcdbd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.475681] env[65385]: DEBUG nova.network.neutron [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance_info_cache with network_info: [{"id": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "address": "fa:16:3e:1a:55:36", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26f849c-40", "ovs_interfaceid": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1016.608735] env[65385]: INFO nova.compute.manager [-] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Took 1.26 seconds to deallocate network for instance. [ 1016.685562] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454213, 'name': ReconfigVM_Task, 'duration_secs': 0.6549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.687740] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Reconfigured VM instance instance-0000005b to attach disk [datastore2] b4bca2c8-fc8f-471a-899b-aac648fee9fe/b4bca2c8-fc8f-471a-899b-aac648fee9fe.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.688683] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6af2d1f-8ba4-47d5-b43e-531291d547c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.697382] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1016.697382] env[65385]: value = "task-4454214" [ 1016.697382] env[65385]: _type = "Task" [ 1016.697382] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.708077] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454214, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.723687] env[65385]: DEBUG nova.compute.manager [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Received event network-vif-plugged-f26f849c-4075-4ee1-b174-88ce8e66b69c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1016.723687] env[65385]: DEBUG oslo_concurrency.lockutils [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Acquiring lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.723687] env[65385]: DEBUG oslo_concurrency.lockutils [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.723687] env[65385]: DEBUG oslo_concurrency.lockutils [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.723687] env[65385]: DEBUG nova.compute.manager [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] No waiting events found dispatching network-vif-plugged-f26f849c-4075-4ee1-b174-88ce8e66b69c {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1016.724332] env[65385]: WARNING nova.compute.manager [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Received unexpected event network-vif-plugged-f26f849c-4075-4ee1-b174-88ce8e66b69c for instance with vm_state building and task_state spawning. [ 1016.724332] env[65385]: DEBUG nova.compute.manager [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Received event network-changed-f26f849c-4075-4ee1-b174-88ce8e66b69c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1016.724332] env[65385]: DEBUG nova.compute.manager [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Refreshing instance network info cache due to event network-changed-f26f849c-4075-4ee1-b174-88ce8e66b69c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1016.724529] env[65385]: DEBUG oslo_concurrency.lockutils [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Acquiring lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.799358] env[65385]: DEBUG nova.scheduler.client.report [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1016.980264] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1016.980741] env[65385]: DEBUG nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Instance network_info: |[{"id": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "address": "fa:16:3e:1a:55:36", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26f849c-40", "ovs_interfaceid": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1016.981464] env[65385]: DEBUG oslo_concurrency.lockutils [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Acquired lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1016.981464] env[65385]: DEBUG nova.network.neutron [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Refreshing network info cache for port f26f849c-4075-4ee1-b174-88ce8e66b69c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1016.983410] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:55:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f26f849c-4075-4ee1-b174-88ce8e66b69c', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.992965] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1016.996601] env[65385]: WARNING neutronclient.v2_0.client [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.997379] env[65385]: WARNING openstack [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.997798] env[65385]: WARNING openstack [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.005748] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.006523] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31e5cd83-e772-4a49-8f1d-b16c2f6a10ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.025503] env[65385]: DEBUG nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1017.029703] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.029703] env[65385]: value = "task-4454215" [ 1017.029703] env[65385]: _type = "Task" [ 1017.029703] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.042439] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454215, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.054522] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1017.054758] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1017.054868] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1017.055055] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1017.055196] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1017.055332] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1017.055597] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.055753] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1017.055939] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1017.056107] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1017.056394] env[65385]: DEBUG nova.virt.hardware [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1017.057697] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b83c7ad-be99-43a5-b65e-d6faa0b9c0ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.069279] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3f1429-b4ca-48b6-8f8b-fbeedf470ab2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.112660] env[65385]: WARNING openstack [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.113128] env[65385]: WARNING openstack [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.121198] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1017.183827] env[65385]: WARNING neutronclient.v2_0.client [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.184452] env[65385]: WARNING openstack [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.184452] env[65385]: WARNING openstack [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.212644] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454214, 'name': Rename_Task, 'duration_secs': 0.178627} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.214576] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.214928] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2eedb78-629d-4370-a533-c10c9a2bb586 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.223785] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1017.223785] env[65385]: value = "task-4454216" [ 1017.223785] env[65385]: _type = "Task" [ 1017.223785] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.241679] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.282380] env[65385]: DEBUG nova.network.neutron [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updated VIF entry in instance network info cache for port f26f849c-4075-4ee1-b174-88ce8e66b69c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1017.282876] env[65385]: DEBUG nova.network.neutron [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance_info_cache with network_info: [{"id": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "address": "fa:16:3e:1a:55:36", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26f849c-40", "ovs_interfaceid": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.305808] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.306393] env[65385]: DEBUG nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1017.309704] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.574s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1017.311264] env[65385]: INFO nova.compute.claims [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.470606] env[65385]: DEBUG nova.network.neutron [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Successfully updated port: a7e0f7ab-d411-40cb-ae92-d58b922f22fe {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1017.541292] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454215, 'name': CreateVM_Task, 'duration_secs': 0.375609} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.541476] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.542089] env[65385]: WARNING neutronclient.v2_0.client [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.542406] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.542565] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.542878] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1017.543150] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f94160ee-f614-4700-bb00-0b9241dc112a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.549081] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1017.549081] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e2b8a-64e5-d531-22d4-fcd9bb9e1d71" [ 1017.549081] env[65385]: _type = "Task" [ 1017.549081] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.557813] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e2b8a-64e5-d531-22d4-fcd9bb9e1d71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.644595] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Completed reading data from the image iterator. {{(pid=65385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1017.644595] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525c588e-814b-4dbb-a07f-e0ca1506bd88/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1017.645626] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fefd952-2c39-4623-8162-a10be5e9b839 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.653087] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525c588e-814b-4dbb-a07f-e0ca1506bd88/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1017.653263] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525c588e-814b-4dbb-a07f-e0ca1506bd88/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1017.653563] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e6ff6aeb-ac5f-48b6-801b-faf2a3085bcb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.735562] env[65385]: DEBUG oslo_vmware.api [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454216, 'name': PowerOnVM_Task, 'duration_secs': 0.505731} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.735882] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.736146] env[65385]: INFO nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Took 8.15 seconds to spawn the instance on the hypervisor. [ 1017.736342] env[65385]: DEBUG nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1017.737242] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a0d07b-2a41-4461-b32d-0ccddd876c8d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.787994] env[65385]: DEBUG oslo_concurrency.lockutils [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] Releasing lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.788295] env[65385]: DEBUG nova.compute.manager [req-68c1ac60-9695-4bad-8474-a8f967f0df69 req-8a781d79-08db-4a7a-bf09-9d1994588139 service nova] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Received event network-vif-deleted-bbadc19c-78ee-4246-a53d-77f01c5ccea8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1017.815686] env[65385]: DEBUG nova.compute.utils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1017.817683] env[65385]: DEBUG nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1017.817880] env[65385]: DEBUG nova.network.neutron [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1017.818256] env[65385]: WARNING neutronclient.v2_0.client [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.818584] env[65385]: WARNING neutronclient.v2_0.client [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.819172] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.819525] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.883830] env[65385]: DEBUG nova.policy [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03f7c78b16e547c38f24ed2c8a329202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ba61d77cc2d4fe1ba3c03466d5985d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1017.889033] env[65385]: DEBUG oslo_vmware.rw_handles [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525c588e-814b-4dbb-a07f-e0ca1506bd88/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1017.889272] env[65385]: INFO nova.virt.vmwareapi.images [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Downloaded image file data 1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa [ 1017.890171] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8383d07b-155f-43f9-a19c-2ca1153193d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.910255] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b0c95ea-b02b-4474-b406-4407885fd4b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.961982] env[65385]: INFO nova.virt.vmwareapi.images [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] The imported VM was unregistered [ 1017.965331] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Caching image {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1017.965680] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating directory with path [datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.966062] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8fe2cce-4195-4ef5-9cf3-da310a5898f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.972569] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-537393de-ea48-4d34-8e04-95331bbc099d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.972744] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-537393de-ea48-4d34-8e04-95331bbc099d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.972918] env[65385]: DEBUG nova.network.neutron [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1017.979953] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created directory with path [datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.980174] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82/OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82.vmdk to [datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa.vmdk. {{(pid=65385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1017.980446] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-20112c1b-1725-47ac-ad7c-c5dc8a9f5496 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.991991] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1017.991991] env[65385]: value = "task-4454218" [ 1017.991991] env[65385]: _type = "Task" [ 1017.991991] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.001748] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454218, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.060384] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e2b8a-64e5-d531-22d4-fcd9bb9e1d71, 'name': SearchDatastore_Task, 'duration_secs': 0.018956} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.060601] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.060854] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.061122] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.061324] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.061451] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.061735] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f4cde9d8-1284-4f15-95f7-5db93ba834f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.072224] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.072417] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.073225] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f4acc27-99f2-428e-beea-2362f3c520dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.080823] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1018.080823] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524b451e-9173-4fec-2b92-aab2786d63f7" [ 1018.080823] env[65385]: _type = "Task" [ 1018.080823] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.090585] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524b451e-9173-4fec-2b92-aab2786d63f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.240240] env[65385]: DEBUG nova.network.neutron [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Successfully created port: 2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1018.259123] env[65385]: INFO nova.compute.manager [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Took 23.68 seconds to build instance. [ 1018.334888] env[65385]: DEBUG nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1018.476709] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.477224] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.510170] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454218, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.546927] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e24124-f9de-4fd6-9bf4-5db4cb26076a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.557496] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ccdd84-6ca6-4ce0-a0b2-5c8f1b6a9951 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.598505] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67154c0-acee-4955-92d8-17f14b1599fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.607916] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524b451e-9173-4fec-2b92-aab2786d63f7, 'name': SearchDatastore_Task, 'duration_secs': 0.010617} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.611133] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffdfc1af-d9f1-47ed-9c89-7eca3a5e5ccf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.614528] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fb2c2e-43d4-45bf-be6a-5c0ded8abd9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.630430] env[65385]: DEBUG nova.compute.provider_tree [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.633982] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1018.633982] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528b3e39-79f8-3375-152a-9c6a83ab5da0" [ 1018.633982] env[65385]: _type = "Task" [ 1018.633982] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.646349] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528b3e39-79f8-3375-152a-9c6a83ab5da0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.738381] env[65385]: DEBUG nova.network.neutron [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1018.761645] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c84a2924-7344-451d-afb9-979a2a571332 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.195s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.784260] env[65385]: DEBUG nova.compute.manager [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Received event network-changed-c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1018.784490] env[65385]: DEBUG nova.compute.manager [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Refreshing instance network info cache due to event network-changed-c11bab1f-7eb9-434c-8436-c2319951b01f. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1018.784794] env[65385]: DEBUG oslo_concurrency.lockutils [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Acquiring lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.785675] env[65385]: DEBUG oslo_concurrency.lockutils [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Acquired lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.785675] env[65385]: DEBUG nova.network.neutron [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Refreshing network info cache for port c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1018.807679] env[65385]: DEBUG nova.compute.manager [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Received event network-vif-plugged-a7e0f7ab-d411-40cb-ae92-d58b922f22fe {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1018.808332] env[65385]: DEBUG oslo_concurrency.lockutils [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Acquiring lock "537393de-ea48-4d34-8e04-95331bbc099d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.808332] env[65385]: DEBUG oslo_concurrency.lockutils [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Lock "537393de-ea48-4d34-8e04-95331bbc099d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.808657] env[65385]: DEBUG oslo_concurrency.lockutils [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Lock "537393de-ea48-4d34-8e04-95331bbc099d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.808736] env[65385]: DEBUG nova.compute.manager [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] No waiting events found dispatching network-vif-plugged-a7e0f7ab-d411-40cb-ae92-d58b922f22fe {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1018.809025] env[65385]: WARNING nova.compute.manager [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Received unexpected event network-vif-plugged-a7e0f7ab-d411-40cb-ae92-d58b922f22fe for instance with vm_state building and task_state spawning. [ 1018.809224] env[65385]: DEBUG nova.compute.manager [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Received event network-changed-a7e0f7ab-d411-40cb-ae92-d58b922f22fe {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1018.809466] env[65385]: DEBUG nova.compute.manager [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Refreshing instance network info cache due to event network-changed-a7e0f7ab-d411-40cb-ae92-d58b922f22fe. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1018.809646] env[65385]: DEBUG oslo_concurrency.lockutils [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Acquiring lock "refresh_cache-537393de-ea48-4d34-8e04-95331bbc099d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.839883] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.840333] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.008182] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454218, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.037747] env[65385]: WARNING neutronclient.v2_0.client [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.038446] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.038793] env[65385]: WARNING openstack [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.136360] env[65385]: DEBUG nova.scheduler.client.report [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1019.159701] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528b3e39-79f8-3375-152a-9c6a83ab5da0, 'name': SearchDatastore_Task, 'duration_secs': 0.086883} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.159701] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.159701] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8a974e99-a7b1-438e-aace-9ae82352495c/8a974e99-a7b1-438e-aace-9ae82352495c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.159701] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43a797ef-929c-4f6b-ab60-1efed6496ec9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.178124] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1019.178124] env[65385]: value = "task-4454219" [ 1019.178124] env[65385]: _type = "Task" [ 1019.178124] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.190425] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.226841] env[65385]: DEBUG nova.network.neutron [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Updating instance_info_cache with network_info: [{"id": "a7e0f7ab-d411-40cb-ae92-d58b922f22fe", "address": "fa:16:3e:09:43:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7e0f7ab-d4", "ovs_interfaceid": "a7e0f7ab-d411-40cb-ae92-d58b922f22fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.289052] env[65385]: WARNING neutronclient.v2_0.client [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.289748] env[65385]: WARNING openstack [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.290209] env[65385]: WARNING openstack [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.352861] env[65385]: DEBUG nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1019.383354] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1019.383771] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1019.383845] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1019.384098] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1019.384271] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1019.384413] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1019.384620] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1019.384786] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1019.385110] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1019.385281] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1019.385457] env[65385]: DEBUG nova.virt.hardware [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1019.386613] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c015671-c56b-4cf6-a0c7-83d406379032 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.398670] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1477c20-0434-4e0f-b7de-5dae70e9d3fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.418450] env[65385]: WARNING openstack [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.418837] env[65385]: WARNING openstack [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.482341] env[65385]: WARNING neutronclient.v2_0.client [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.483076] env[65385]: WARNING openstack [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.483568] env[65385]: WARNING openstack [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.506458] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454218, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.581212] env[65385]: DEBUG nova.network.neutron [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updated VIF entry in instance network info cache for port c11bab1f-7eb9-434c-8436-c2319951b01f. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1019.581684] env[65385]: DEBUG nova.network.neutron [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updating instance_info_cache with network_info: [{"id": "c11bab1f-7eb9-434c-8436-c2319951b01f", "address": "fa:16:3e:ea:99:f4", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc11bab1f-7e", "ovs_interfaceid": "c11bab1f-7eb9-434c-8436-c2319951b01f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.651180] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.651729] env[65385]: DEBUG nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1019.654576] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.607s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.654859] env[65385]: DEBUG nova.objects.instance [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lazy-loading 'resources' on Instance uuid cb99e6ec-fefc-4f44-86a0-9d886e9c722d {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.691274] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.729964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-537393de-ea48-4d34-8e04-95331bbc099d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.730339] env[65385]: DEBUG nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Instance network_info: |[{"id": "a7e0f7ab-d411-40cb-ae92-d58b922f22fe", "address": "fa:16:3e:09:43:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7e0f7ab-d4", "ovs_interfaceid": "a7e0f7ab-d411-40cb-ae92-d58b922f22fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1019.730680] env[65385]: DEBUG oslo_concurrency.lockutils [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Acquired lock "refresh_cache-537393de-ea48-4d34-8e04-95331bbc099d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1019.730856] env[65385]: DEBUG nova.network.neutron [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Refreshing network info cache for port a7e0f7ab-d411-40cb-ae92-d58b922f22fe {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1019.732367] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:43:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7e0f7ab-d411-40cb-ae92-d58b922f22fe', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.740980] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating folder: Project (94cc5bc221c0455ea760f5022db6bdc4). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1019.742255] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14d6bd5d-d8ef-4812-abdd-54292753966f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.757434] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created folder: Project (94cc5bc221c0455ea760f5022db6bdc4) in parent group-v870881. [ 1019.757649] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating folder: Instances. Parent ref: group-v871146. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1019.757941] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e23b5c0-c4a1-4139-9332-b62e4db67864 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.772435] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created folder: Instances in parent group-v871146. [ 1019.772714] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1019.774659] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1019.774659] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d30052d6-0884-43f1-a165-d5de2228a3ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.797915] env[65385]: DEBUG nova.network.neutron [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Successfully updated port: 2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1019.809209] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.809209] env[65385]: value = "task-4454222" [ 1019.809209] env[65385]: _type = "Task" [ 1019.809209] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.822712] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454222, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.007328] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454218, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.084856] env[65385]: DEBUG oslo_concurrency.lockutils [req-58270eeb-734f-48a5-82df-9a426b3a4881 req-84334476-f22a-489e-b56e-bdff19b57da2 service nova] Releasing lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.160462] env[65385]: DEBUG nova.compute.utils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1020.167149] env[65385]: DEBUG nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1020.167517] env[65385]: DEBUG nova.network.neutron [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1020.167994] env[65385]: WARNING neutronclient.v2_0.client [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.168272] env[65385]: WARNING neutronclient.v2_0.client [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.168873] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.169310] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.194996] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.233506] env[65385]: DEBUG nova.policy [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662871f375e441eb98d814cf7d7ac5c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bc8afc699e34f059f18d8244980bc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1020.245135] env[65385]: WARNING neutronclient.v2_0.client [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.245903] env[65385]: WARNING openstack [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.246305] env[65385]: WARNING openstack [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.301209] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.301414] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.301466] env[65385]: DEBUG nova.network.neutron [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1020.328456] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454222, 'name': CreateVM_Task, 'duration_secs': 0.391475} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.332401] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.333379] env[65385]: WARNING neutronclient.v2_0.client [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.333880] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.334159] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.334556] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1020.334973] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbecc695-2d63-4000-8817-0a5701450ffc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.346225] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1020.346225] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3fd82-c10f-bb92-8187-d2c84f20a9c4" [ 1020.346225] env[65385]: _type = "Task" [ 1020.346225] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.364074] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3fd82-c10f-bb92-8187-d2c84f20a9c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.446835] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a3733a-2a18-41f8-b11a-a72a7a029883 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.456219] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8ece32-b6f1-4c10-b7ed-4ed2243c279b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.517166] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e1756d-795d-4bc3-926e-7f5189128b2f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.527818] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454218, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.472297} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.531059] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82/OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82.vmdk to [datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa.vmdk. [ 1020.531402] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Cleaning up location [datastore1] OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1020.531579] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_3dda3ef2-2d89-447d-adc1-8a7cfd544d82 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.531943] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c458d05-a7bc-49c4-a998-7f1f5a6572ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.536285] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d04684-03da-4de3-8734-e3e75879e36b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.547518] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1020.547518] env[65385]: value = "task-4454223" [ 1020.547518] env[65385]: _type = "Task" [ 1020.547518] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.560072] env[65385]: DEBUG nova.compute.provider_tree [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.573206] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.577852] env[65385]: WARNING openstack [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.578318] env[65385]: WARNING openstack [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.605720] env[65385]: DEBUG nova.network.neutron [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Successfully created port: 62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1020.650184] env[65385]: WARNING neutronclient.v2_0.client [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.650903] env[65385]: WARNING openstack [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.651330] env[65385]: WARNING openstack [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.667452] env[65385]: DEBUG nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1020.694188] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454219, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.750538] env[65385]: DEBUG nova.network.neutron [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Updated VIF entry in instance network info cache for port a7e0f7ab-d411-40cb-ae92-d58b922f22fe. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1020.750979] env[65385]: DEBUG nova.network.neutron [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Updating instance_info_cache with network_info: [{"id": "a7e0f7ab-d411-40cb-ae92-d58b922f22fe", "address": "fa:16:3e:09:43:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7e0f7ab-d4", "ovs_interfaceid": "a7e0f7ab-d411-40cb-ae92-d58b922f22fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1020.804744] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.805175] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.845073] env[65385]: DEBUG nova.network.neutron [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1020.853461] env[65385]: DEBUG nova.compute.manager [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Received event network-vif-plugged-2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1020.853675] env[65385]: DEBUG oslo_concurrency.lockutils [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Acquiring lock "5470afe7-6dd0-4778-9b83-2956ede04b7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1020.853966] env[65385]: DEBUG oslo_concurrency.lockutils [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1020.854246] env[65385]: DEBUG oslo_concurrency.lockutils [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.854421] env[65385]: DEBUG nova.compute.manager [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] No waiting events found dispatching network-vif-plugged-2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1020.854788] env[65385]: WARNING nova.compute.manager [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Received unexpected event network-vif-plugged-2ea48afc-dea7-435d-996b-e1f73fcec4bf for instance with vm_state building and task_state spawning. [ 1020.854788] env[65385]: DEBUG nova.compute.manager [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Received event network-changed-2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1020.854915] env[65385]: DEBUG nova.compute.manager [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Refreshing instance network info cache due to event network-changed-2ea48afc-dea7-435d-996b-e1f73fcec4bf. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1020.855013] env[65385]: DEBUG oslo_concurrency.lockutils [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Acquiring lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.863700] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c3fd82-c10f-bb92-8187-d2c84f20a9c4, 'name': SearchDatastore_Task, 'duration_secs': 0.056839} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.864053] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.864290] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.864511] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.864650] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.864824] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.866239] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.866683] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.873095] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95c824bb-e98f-4372-ac88-d0ec17deb1bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.884408] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.884676] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1020.885469] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5685aa8a-f625-46ec-8303-f27c9d20b92e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.891867] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1020.891867] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fe5ca-edb7-d079-f44d-5e69c467e1b4" [ 1020.891867] env[65385]: _type = "Task" [ 1020.891867] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.906112] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fe5ca-edb7-d079-f44d-5e69c467e1b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.933101] env[65385]: WARNING neutronclient.v2_0.client [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.933827] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.934276] env[65385]: WARNING openstack [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.013777] env[65385]: DEBUG nova.network.neutron [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updating instance_info_cache with network_info: [{"id": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "address": "fa:16:3e:00:05:77", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea48afc-de", "ovs_interfaceid": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1021.067409] env[65385]: DEBUG nova.scheduler.client.report [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1021.074144] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211423} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.074346] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.074491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.074721] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa.vmdk to [datastore1] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1021.074979] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-847bd24e-e860-4332-abf8-16d2f6c14070 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.083063] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1021.083063] env[65385]: value = "task-4454224" [ 1021.083063] env[65385]: _type = "Task" [ 1021.083063] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.093635] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454224, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.189868] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454219, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.61579} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.190065] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 8a974e99-a7b1-438e-aace-9ae82352495c/8a974e99-a7b1-438e-aace-9ae82352495c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1021.190494] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.190571] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-345d9f6a-280a-483d-acdb-6f29cb887676 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.198944] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1021.198944] env[65385]: value = "task-4454225" [ 1021.198944] env[65385]: _type = "Task" [ 1021.198944] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.209460] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454225, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.254745] env[65385]: DEBUG oslo_concurrency.lockutils [req-d05adaeb-ce6e-4bee-a12e-3b80724976ae req-9f29b7ae-7341-4c2b-8f06-3abe3b5caf39 service nova] Releasing lock "refresh_cache-537393de-ea48-4d34-8e04-95331bbc099d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.403683] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529fe5ca-edb7-d079-f44d-5e69c467e1b4, 'name': SearchDatastore_Task, 'duration_secs': 0.010786} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.404655] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28f32f74-4f8e-45cd-8525-1038ab6b3136 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.411644] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1021.411644] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bf9af-ce19-5fdd-2278-078277c9781f" [ 1021.411644] env[65385]: _type = "Task" [ 1021.411644] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.423193] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bf9af-ce19-5fdd-2278-078277c9781f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.516946] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.517720] env[65385]: DEBUG nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Instance network_info: |[{"id": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "address": "fa:16:3e:00:05:77", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea48afc-de", "ovs_interfaceid": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1021.518357] env[65385]: DEBUG oslo_concurrency.lockutils [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Acquired lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.518677] env[65385]: DEBUG nova.network.neutron [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Refreshing network info cache for port 2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1021.520587] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:05:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ea48afc-dea7-435d-996b-e1f73fcec4bf', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1021.528933] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1021.530126] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1021.530419] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19359c28-96a2-43f8-823b-b539bf871d11 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.559887] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1021.559887] env[65385]: value = "task-4454226" [ 1021.559887] env[65385]: _type = "Task" [ 1021.559887] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.572601] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454226, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.575639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.921s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.578127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.043s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.578397] env[65385]: DEBUG nova.objects.instance [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lazy-loading 'resources' on Instance uuid 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.598644] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454224, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.609855] env[65385]: INFO nova.scheduler.client.report [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted allocations for instance cb99e6ec-fefc-4f44-86a0-9d886e9c722d [ 1021.678692] env[65385]: DEBUG nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1021.713832] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1021.714215] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1021.714510] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1021.714777] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1021.715038] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1021.715293] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1021.715544] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.715738] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1021.715943] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1021.716136] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1021.716444] env[65385]: DEBUG nova.virt.hardware [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1021.717402] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60352b61-8a7f-4a86-9f6c-4488ea43b066 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.724499] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454225, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.236332} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.725345] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.726190] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c510c2-b20a-421c-adc0-1d82ed285bce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.735772] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456d505d-0e7f-4b39-8a61-fea5dac3d34c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.760896] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 8a974e99-a7b1-438e-aace-9ae82352495c/8a974e99-a7b1-438e-aace-9ae82352495c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.761795] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3e501bb-140f-473c-86c1-2736931f6323 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.792389] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1021.792389] env[65385]: value = "task-4454227" [ 1021.792389] env[65385]: _type = "Task" [ 1021.792389] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.803782] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454227, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.925278] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bf9af-ce19-5fdd-2278-078277c9781f, 'name': SearchDatastore_Task, 'duration_secs': 0.091807} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.925678] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.925965] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1021.926253] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6824fb5a-aa8e-4624-a1f0-f6f2c8ceb480 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.935430] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1021.935430] env[65385]: value = "task-4454228" [ 1021.935430] env[65385]: _type = "Task" [ 1021.935430] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.946855] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.031140] env[65385]: WARNING neutronclient.v2_0.client [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1022.031967] env[65385]: WARNING openstack [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.032364] env[65385]: WARNING openstack [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.072930] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454226, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.100140] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454224, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.124837] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e55778a-0a6f-4dd6-a6d3-4d761bebf576 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "cb99e6ec-fefc-4f44-86a0-9d886e9c722d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.493s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.145439] env[65385]: DEBUG nova.network.neutron [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Successfully updated port: 62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1022.157521] env[65385]: WARNING openstack [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.157707] env[65385]: WARNING openstack [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.237246] env[65385]: WARNING neutronclient.v2_0.client [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1022.237946] env[65385]: WARNING openstack [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.238382] env[65385]: WARNING openstack [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.305035] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5753d09a-7f7c-4b1f-8722-fc7b2c1ed49a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.312011] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454227, 'name': ReconfigVM_Task, 'duration_secs': 0.393446} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.314481] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 8a974e99-a7b1-438e-aace-9ae82352495c/8a974e99-a7b1-438e-aace-9ae82352495c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.315352] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1eaef273-5a5a-4d26-9828-49adfce78125 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.318043] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0295f3aa-db8f-47b3-b837-c071276fde97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.329198] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1022.329198] env[65385]: value = "task-4454229" [ 1022.329198] env[65385]: _type = "Task" [ 1022.329198] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.360074] env[65385]: DEBUG nova.network.neutron [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updated VIF entry in instance network info cache for port 2ea48afc-dea7-435d-996b-e1f73fcec4bf. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1022.360538] env[65385]: DEBUG nova.network.neutron [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updating instance_info_cache with network_info: [{"id": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "address": "fa:16:3e:00:05:77", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea48afc-de", "ovs_interfaceid": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1022.366645] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7b613c-eb61-45c0-a07b-fd02c57eca8f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.380287] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f648a8-401c-4229-95e5-1fff00f6cf16 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.385365] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454229, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.400435] env[65385]: DEBUG nova.compute.provider_tree [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.449809] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.577097] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454226, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.598943] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454224, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.650621] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "refresh_cache-992920c3-af4f-489e-b21f-9d52c24399de" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.650817] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "refresh_cache-992920c3-af4f-489e-b21f-9d52c24399de" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.651010] env[65385]: DEBUG nova.network.neutron [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1022.870036] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454229, 'name': Rename_Task, 'duration_secs': 0.208501} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.870592] env[65385]: DEBUG oslo_concurrency.lockutils [req-373c6154-44a5-4943-ab85-503fd8d94150 req-663f7169-f67d-4d97-a8e3-88ae9b847fe8 service nova] Releasing lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.871100] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.871371] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1edc994-a236-417c-910d-e5b22dbc0b54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.883561] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1022.883561] env[65385]: value = "task-4454230" [ 1022.883561] env[65385]: _type = "Task" [ 1022.883561] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.890964] env[65385]: DEBUG nova.compute.manager [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Received event network-vif-plugged-62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1022.891162] env[65385]: DEBUG oslo_concurrency.lockutils [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Acquiring lock "992920c3-af4f-489e-b21f-9d52c24399de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1022.891397] env[65385]: DEBUG oslo_concurrency.lockutils [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Lock "992920c3-af4f-489e-b21f-9d52c24399de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1022.891591] env[65385]: DEBUG oslo_concurrency.lockutils [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Lock "992920c3-af4f-489e-b21f-9d52c24399de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.891835] env[65385]: DEBUG nova.compute.manager [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] No waiting events found dispatching network-vif-plugged-62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1022.892093] env[65385]: WARNING nova.compute.manager [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Received unexpected event network-vif-plugged-62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 for instance with vm_state building and task_state spawning. [ 1022.892242] env[65385]: DEBUG nova.compute.manager [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Received event network-changed-62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1022.892471] env[65385]: DEBUG nova.compute.manager [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Refreshing instance network info cache due to event network-changed-62dbbb11-2d9c-4536-a2ed-e07b1e51dff7. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1022.892646] env[65385]: DEBUG oslo_concurrency.lockutils [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Acquiring lock "refresh_cache-992920c3-af4f-489e-b21f-9d52c24399de" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.896873] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454230, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.903638] env[65385]: DEBUG nova.scheduler.client.report [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1022.949555] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.074500] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454226, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.099486] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454224, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.154902] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.155501] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.198948] env[65385]: DEBUG nova.network.neutron [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1023.224923] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.225347] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.298634] env[65385]: WARNING neutronclient.v2_0.client [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.299948] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.300545] env[65385]: WARNING openstack [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.398027] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454230, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.409157] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.411824] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.004s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.412113] env[65385]: DEBUG nova.objects.instance [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'resources' on Instance uuid 370b6aa8-2337-48e1-8102-611ce4494a55 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.418058] env[65385]: DEBUG nova.network.neutron [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Updating instance_info_cache with network_info: [{"id": "62dbbb11-2d9c-4536-a2ed-e07b1e51dff7", "address": "fa:16:3e:81:16:bc", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62dbbb11-2d", "ovs_interfaceid": "62dbbb11-2d9c-4536-a2ed-e07b1e51dff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1023.443059] env[65385]: INFO nova.scheduler.client.report [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Deleted allocations for instance 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5 [ 1023.450575] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.572454] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454226, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.600039] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454224, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.468392} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.600039] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa/1eae85d6-3c5c-444e-a00b-66cf0e1c4aaa.vmdk to [datastore1] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.600704] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f81ea4-fccd-425a-8e7f-6d9ec61847ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.627063] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.627481] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f59e880e-0d6e-41c3-9fdd-89588d3fa21c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.653048] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1023.653048] env[65385]: value = "task-4454231" [ 1023.653048] env[65385]: _type = "Task" [ 1023.653048] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.663522] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454231, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.894808] env[65385]: DEBUG oslo_vmware.api [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454230, 'name': PowerOnVM_Task, 'duration_secs': 0.520416} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.895245] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.895468] env[65385]: INFO nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Took 9.65 seconds to spawn the instance on the hypervisor. [ 1023.895646] env[65385]: DEBUG nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1023.896431] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716acf32-3d9f-446b-a702-83dcea5c3e86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.920984] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "refresh_cache-992920c3-af4f-489e-b21f-9d52c24399de" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.921416] env[65385]: DEBUG nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Instance network_info: |[{"id": "62dbbb11-2d9c-4536-a2ed-e07b1e51dff7", "address": "fa:16:3e:81:16:bc", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62dbbb11-2d", "ovs_interfaceid": "62dbbb11-2d9c-4536-a2ed-e07b1e51dff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1023.921760] env[65385]: DEBUG oslo_concurrency.lockutils [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Acquired lock "refresh_cache-992920c3-af4f-489e-b21f-9d52c24399de" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.921934] env[65385]: DEBUG nova.network.neutron [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Refreshing network info cache for port 62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1023.923103] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:16:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62dbbb11-2d9c-4536-a2ed-e07b1e51dff7', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.930652] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating folder: Project (6bc8afc699e34f059f18d8244980bc1d). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1023.931762] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7afaa4e0-1185-40e8-98b9-7b939623e7b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.949032] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454228, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.936334} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.950446] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.950687] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.950951] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created folder: Project (6bc8afc699e34f059f18d8244980bc1d) in parent group-v870881. [ 1023.951127] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating folder: Instances. Parent ref: group-v871150. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1023.954787] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a255a8d5-82ff-4f51-91f9-dce73a4dae50 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.958799] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0692d287-a25e-4c2d-9458-4dc6b05cff49 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.961223] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6456d81b-8bc1-4f3c-a12a-96ac71b00e66 tempest-MultipleCreateTestJSON-115352278 tempest-MultipleCreateTestJSON-115352278-project-member] Lock "6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.183s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.970194] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1023.970194] env[65385]: value = "task-4454234" [ 1023.970194] env[65385]: _type = "Task" [ 1023.970194] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.977553] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created folder: Instances in parent group-v871150. [ 1023.977860] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1023.979025] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1023.980195] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37a93a13-d853-40d6-aff2-99799ddf9af2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.001796] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.007303] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.007303] env[65385]: value = "task-4454235" [ 1024.007303] env[65385]: _type = "Task" [ 1024.007303] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.018766] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454235, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.076065] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454226, 'name': CreateVM_Task, 'duration_secs': 2.146201} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.076206] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1024.076733] env[65385]: WARNING neutronclient.v2_0.client [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.077094] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.077255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1024.077560] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1024.077814] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e0e53b4-ce04-41c9-ad2f-1203a345e9f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.086129] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1024.086129] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523beeb3-3c08-4a68-95a6-8cb3c08c3372" [ 1024.086129] env[65385]: _type = "Task" [ 1024.086129] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.096618] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523beeb3-3c08-4a68-95a6-8cb3c08c3372, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.108027] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125c1ebc-a898-4998-917e-ac930af138a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.114744] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83c5092-1eea-4da2-9d94-ba579a2d0b3b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.151515] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98969ebd-b572-48ac-9aac-ab392fa4b69c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.166688] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5990dd8-8e45-4e4e-a259-6611c8ebe845 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.170725] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454231, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.182354] env[65385]: DEBUG nova.compute.provider_tree [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.414877] env[65385]: INFO nova.compute.manager [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Took 28.14 seconds to build instance. [ 1024.431906] env[65385]: WARNING neutronclient.v2_0.client [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.432610] env[65385]: WARNING openstack [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.432964] env[65385]: WARNING openstack [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.482562] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454234, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073285} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.483153] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.483636] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764a5bb0-c1b4-4149-ab69-738fae967255 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.509435] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.510239] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55381fa3-a405-4f6a-9db4-ff53be414c18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.543615] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454235, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.545713] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1024.545713] env[65385]: value = "task-4454236" [ 1024.545713] env[65385]: _type = "Task" [ 1024.545713] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.557043] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454236, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.579884] env[65385]: WARNING openstack [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.580490] env[65385]: WARNING openstack [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.603697] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523beeb3-3c08-4a68-95a6-8cb3c08c3372, 'name': SearchDatastore_Task, 'duration_secs': 0.010051} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.604053] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.604425] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.604800] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.605064] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1024.605374] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.605829] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c34159b9-be94-449d-8f5a-3d4da0f97b53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.621090] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.621090] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1024.621090] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6354c14-62d8-47f6-b2a0-8a39f4e24914 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.628681] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1024.628681] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52460839-9410-84ed-d53f-9ca046493d2a" [ 1024.628681] env[65385]: _type = "Task" [ 1024.628681] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.641496] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52460839-9410-84ed-d53f-9ca046493d2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.669132] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454231, 'name': ReconfigVM_Task, 'duration_secs': 0.953084} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.669488] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7/048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.670935] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'boot_index': 0, 'encryption_secret_uuid': None, 'device_type': 'disk', 'guest_format': None, 'encryption_options': None, 'encrypted': False, 'encryption_format': None, 'device_name': '/dev/sda', 'disk_bus': None, 'size': 0, 'image_id': 'c3e9f50a-a10e-4952-ab03-d986996039fa'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871137', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'name': 'volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'serial': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2'}, 'device_type': None, 'guest_format': None, 'attachment_id': '8486e3ce-0fa0-465e-b22d-70dc7d162e26', 'mount_device': '/dev/sdb', 'disk_bus': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=65385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1024.671156] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1024.671343] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871137', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'name': 'volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'serial': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1024.672199] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76605160-d885-4ea3-aee7-bc4f891422f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.690051] env[65385]: DEBUG nova.scheduler.client.report [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1024.695624] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b7c695-47c1-41e9-8557-338a63dee64d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.725293] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2/volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.726693] env[65385]: WARNING neutronclient.v2_0.client [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.727407] env[65385]: WARNING openstack [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.727763] env[65385]: WARNING openstack [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.735724] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6a6b943-59ee-47d2-a691-bba125f8653d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.758112] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1024.758112] env[65385]: value = "task-4454237" [ 1024.758112] env[65385]: _type = "Task" [ 1024.758112] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.767533] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454237, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.842502] env[65385]: DEBUG nova.network.neutron [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Updated VIF entry in instance network info cache for port 62dbbb11-2d9c-4536-a2ed-e07b1e51dff7. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1024.842875] env[65385]: DEBUG nova.network.neutron [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Updating instance_info_cache with network_info: [{"id": "62dbbb11-2d9c-4536-a2ed-e07b1e51dff7", "address": "fa:16:3e:81:16:bc", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62dbbb11-2d", "ovs_interfaceid": "62dbbb11-2d9c-4536-a2ed-e07b1e51dff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1024.917986] env[65385]: DEBUG oslo_concurrency.lockutils [None req-83fc6af4-45aa-4b4d-b43b-e8a04468fd85 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.652s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.022197] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454235, 'name': CreateVM_Task, 'duration_secs': 0.920631} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.022384] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1025.022901] env[65385]: WARNING neutronclient.v2_0.client [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.023298] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.023450] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1025.023750] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1025.024117] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b6cf99e-33f6-47a8-8ad0-c1dc329166ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.029763] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1025.029763] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c0356-8440-90c6-ec2a-aabf25af95e2" [ 1025.029763] env[65385]: _type = "Task" [ 1025.029763] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.039864] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c0356-8440-90c6-ec2a-aabf25af95e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.057322] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454236, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.141102] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52460839-9410-84ed-d53f-9ca046493d2a, 'name': SearchDatastore_Task, 'duration_secs': 0.011895} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.141939] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9c14829-c29b-450d-8aa7-57cb6ee4459c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.149102] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1025.149102] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5241b643-bb79-7a41-bba3-ab07eeccb2d3" [ 1025.149102] env[65385]: _type = "Task" [ 1025.149102] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.158213] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5241b643-bb79-7a41-bba3-ab07eeccb2d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.196340] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.199040] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.120s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.215624] env[65385]: INFO nova.scheduler.client.report [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted allocations for instance 370b6aa8-2337-48e1-8102-611ce4494a55 [ 1025.269405] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.346817] env[65385]: DEBUG oslo_concurrency.lockutils [req-174359b7-7c02-45f8-8172-4602f6aea6d3 req-ce1d02bb-af17-415a-86a0-c1fb44e0ab87 service nova] Releasing lock "refresh_cache-992920c3-af4f-489e-b21f-9d52c24399de" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1025.545121] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c0356-8440-90c6-ec2a-aabf25af95e2, 'name': SearchDatastore_Task, 'duration_secs': 0.012348} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.545812] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1025.545812] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.546083] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.546231] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1025.546485] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.546971] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d24706b-f2c9-42eb-b545-6df345323349 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.557883] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454236, 'name': ReconfigVM_Task, 'duration_secs': 0.883718} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.558197] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.559800] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fea2ac7b-a7e2-4c4c-8896-0aee3524421e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.561709] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.561882] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1025.562708] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05024756-a1f3-4812-93cc-1e1e1f7a4344 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.568926] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1025.568926] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5218eec8-3b8b-2f81-41ad-cf2ae4373570" [ 1025.568926] env[65385]: _type = "Task" [ 1025.568926] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.573931] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1025.573931] env[65385]: value = "task-4454238" [ 1025.573931] env[65385]: _type = "Task" [ 1025.573931] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.580709] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5218eec8-3b8b-2f81-41ad-cf2ae4373570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.587467] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454238, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.660706] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5241b643-bb79-7a41-bba3-ab07eeccb2d3, 'name': SearchDatastore_Task, 'duration_secs': 0.019422} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.660988] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1025.661269] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5470afe7-6dd0-4778-9b83-2956ede04b7f/5470afe7-6dd0-4778-9b83-2956ede04b7f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1025.661542] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08b315fe-1d47-4939-b69a-4f3d47ed0451 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.670654] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1025.670654] env[65385]: value = "task-4454239" [ 1025.670654] env[65385]: _type = "Task" [ 1025.670654] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.681233] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.727574] env[65385]: DEBUG oslo_concurrency.lockutils [None req-69ddccf2-5611-4368-815b-6a78687cb42a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "370b6aa8-2337-48e1-8102-611ce4494a55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.138s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.769054] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454237, 'name': ReconfigVM_Task, 'duration_secs': 0.845733} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.769341] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfigured VM instance instance-00000049 to attach disk [datastore1] volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2/volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.774197] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee88a5c1-56f5-4720-bee7-87efebdb0d33 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.791459] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1025.791459] env[65385]: value = "task-4454240" [ 1025.791459] env[65385]: _type = "Task" [ 1025.791459] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.803142] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454240, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.084339] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5218eec8-3b8b-2f81-41ad-cf2ae4373570, 'name': SearchDatastore_Task, 'duration_secs': 0.010109} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.085868] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2384033d-f037-407a-88f4-31ea1ee46300 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.092827] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454238, 'name': Rename_Task, 'duration_secs': 0.183756} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.092827] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.093150] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d432c35d-bb29-478e-ad54-53b475136fb4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.098500] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1026.098500] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d5ca16-001c-041e-d600-4547e209c9f8" [ 1026.098500] env[65385]: _type = "Task" [ 1026.098500] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.104252] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1026.104252] env[65385]: value = "task-4454241" [ 1026.104252] env[65385]: _type = "Task" [ 1026.104252] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.114585] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d5ca16-001c-041e-d600-4547e209c9f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.124733] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454241, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.150990] env[65385]: DEBUG nova.compute.manager [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1026.186226] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454239, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502416} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.186548] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5470afe7-6dd0-4778-9b83-2956ede04b7f/5470afe7-6dd0-4778-9b83-2956ede04b7f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1026.186771] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.187079] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-249ec85c-f68d-4fc7-b623-b75658c8e5f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.198034] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1026.198034] env[65385]: value = "task-4454242" [ 1026.198034] env[65385]: _type = "Task" [ 1026.198034] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.211576] env[65385]: INFO nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating resource usage from migration 64929b85-d30a-41d5-9d93-bceef0b8938f [ 1026.214390] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454242, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.254149] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 59f26752-75ab-45e5-b0a9-bbcef44ecd7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.254362] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.254593] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 0c345794-036c-45d1-985e-cf5c116bd873 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1026.254900] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance a5fe2ad9-6dfe-485b-bf27-beb218d31467 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.254900] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.255051] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b4bca2c8-fc8f-471a-899b-aac648fee9fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.255188] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 537393de-ea48-4d34-8e04-95331bbc099d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.255326] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 5470afe7-6dd0-4778-9b83-2956ede04b7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.255481] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 992920c3-af4f-489e-b21f-9d52c24399de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.255618] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Migration 64929b85-d30a-41d5-9d93-bceef0b8938f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1026.256161] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 8a974e99-a7b1-438e-aace-9ae82352495c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1026.256464] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1026.256644] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=100GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '9', 'num_vm_active': '5', 'num_task_None': '4', 'num_os_type_None': '9', 'num_proj_60de0d1c162342209795a180391661f6': '1', 'io_workload': '4', 'num_vm_shelved_offloaded': '1', 'num_task_spawning': '4', 'num_proj_b27c092a0fd345ea953956469c21f8bb': '1', 'num_proj_1598307c304d46a884a88b3eb5f5d7dc': '1', 'num_proj_1991528775bd4557ae6150b30095e923': '1', 'num_proj_8d45ff3446a84ef2a1945f7a66b747e8': '1', 'num_task_resize_prep': '1', 'num_proj_754b1facaaa14501b2204c98e1d7a5a8': '1', 'num_vm_building': '3', 'num_proj_94cc5bc221c0455ea760f5022db6bdc4': '1', 'num_proj_2ba61d77cc2d4fe1ba3c03466d5985d9': '1', 'num_proj_6bc8afc699e34f059f18d8244980bc1d': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1026.306226] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454240, 'name': ReconfigVM_Task, 'duration_secs': 0.19214} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.306721] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871137', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'name': 'volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'serial': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1026.307578] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ffeb71f-2dbc-4068-953e-307a24aa05fd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.317602] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1026.317602] env[65385]: value = "task-4454243" [ 1026.317602] env[65385]: _type = "Task" [ 1026.317602] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.327952] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454243, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.448529] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b51a25-d465-4907-b441-58e8cde5edc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.456824] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276ba9dd-10f8-47ea-812b-f272f6cec20c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.490615] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb6dd4b-911c-4ac5-a22b-30f94bbbfc0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.499170] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486a17b1-b563-4eaa-a000-34ac253bfcda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.517342] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.621017] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d5ca16-001c-041e-d600-4547e209c9f8, 'name': SearchDatastore_Task, 'duration_secs': 0.037062} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.621268] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454241, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.621521] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1026.621865] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 992920c3-af4f-489e-b21f-9d52c24399de/992920c3-af4f-489e-b21f-9d52c24399de.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1026.622073] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5b0559b-3081-40b9-ac8a-f1129db2d9c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.630318] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1026.630318] env[65385]: value = "task-4454244" [ 1026.630318] env[65385]: _type = "Task" [ 1026.630318] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.642787] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454244, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.676662] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.709273] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454242, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073136} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.709558] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.710392] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7539766-f572-408e-b188-f83a343a9811 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.734801] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 5470afe7-6dd0-4778-9b83-2956ede04b7f/5470afe7-6dd0-4778-9b83-2956ede04b7f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.735187] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d704b3c5-c189-4e0a-b719-b0f66667cab9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.757712] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1026.757712] env[65385]: value = "task-4454245" [ 1026.757712] env[65385]: _type = "Task" [ 1026.757712] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.767497] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454245, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.834072] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454243, 'name': Rename_Task, 'duration_secs': 0.334985} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.834421] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.834689] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29f54cea-e3b9-4d0f-856b-c8aa19bc6850 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.845528] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1026.845528] env[65385]: value = "task-4454246" [ 1026.845528] env[65385]: _type = "Task" [ 1026.845528] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.858579] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454246, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.022551] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1027.121463] env[65385]: DEBUG oslo_vmware.api [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454241, 'name': PowerOnVM_Task, 'duration_secs': 0.853055} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.121745] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1027.121958] env[65385]: INFO nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Took 10.10 seconds to spawn the instance on the hypervisor. [ 1027.122256] env[65385]: DEBUG nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1027.123165] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24811c86-8689-406a-8678-8e52a7f8a850 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.145758] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454244, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.274462] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454245, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.356808] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454246, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.530837] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1027.531540] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.332s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.531540] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.410s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.531540] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.533601] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.857s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.555407] env[65385]: INFO nova.scheduler.client.report [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Deleted allocations for instance 0c345794-036c-45d1-985e-cf5c116bd873 [ 1027.645713] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454244, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519084} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.646372] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 992920c3-af4f-489e-b21f-9d52c24399de/992920c3-af4f-489e-b21f-9d52c24399de.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1027.646480] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.648697] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d952b97-0be4-4dfd-98da-8fc8dab2e327 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.651323] env[65385]: INFO nova.compute.manager [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Took 29.89 seconds to build instance. [ 1027.658895] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1027.658895] env[65385]: value = "task-4454247" [ 1027.658895] env[65385]: _type = "Task" [ 1027.658895] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.669397] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454247, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.770492] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454245, 'name': ReconfigVM_Task, 'duration_secs': 0.572846} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.770809] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 5470afe7-6dd0-4778-9b83-2956ede04b7f/5470afe7-6dd0-4778-9b83-2956ede04b7f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.771503] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-781f1183-ee14-46d3-8f3c-e6528b0add59 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.780250] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1027.780250] env[65385]: value = "task-4454248" [ 1027.780250] env[65385]: _type = "Task" [ 1027.780250] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.790761] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454248, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.858234] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454246, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.934838] env[65385]: DEBUG nova.compute.manager [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1028.042964] env[65385]: INFO nova.compute.claims [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.063700] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f3d807a7-1f92-446f-aac6-4776b5b505e0 tempest-ServersNegativeTestJSON-537966175 tempest-ServersNegativeTestJSON-537966175-project-member] Lock "0c345794-036c-45d1-985e-cf5c116bd873" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.420s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.124998] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.125464] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1028.125820] env[65385]: DEBUG nova.objects.instance [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'flavor' on Instance uuid a5fe2ad9-6dfe-485b-bf27-beb218d31467 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.153891] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a714ac3-90a2-44fb-9233-d29bb706c7bb tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "537393de-ea48-4d34-8e04-95331bbc099d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.398s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.171410] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454247, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073875} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.171625] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.172350] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fc818c-78ba-4102-9506-1d60aed985e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.197473] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 992920c3-af4f-489e-b21f-9d52c24399de/992920c3-af4f-489e-b21f-9d52c24399de.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.197806] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e3d1927-23f0-4170-ac69-6ec579d907c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.218721] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1028.218721] env[65385]: value = "task-4454249" [ 1028.218721] env[65385]: _type = "Task" [ 1028.218721] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.230244] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454249, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.290514] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454248, 'name': Rename_Task, 'duration_secs': 0.167984} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.290736] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1028.290936] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81bdc464-a81f-45f2-a1ad-9ae4ae68cf7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.298637] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1028.298637] env[65385]: value = "task-4454250" [ 1028.298637] env[65385]: _type = "Task" [ 1028.298637] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.307991] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.357920] env[65385]: DEBUG oslo_vmware.api [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454246, 'name': PowerOnVM_Task, 'duration_secs': 1.154288} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.358417] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.475097] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.529039] env[65385]: DEBUG nova.compute.manager [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1028.529039] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42087dc9-3815-47ba-b50a-6e8bd5bad69a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.550988] env[65385]: INFO nova.compute.resource_tracker [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating resource usage from migration 64929b85-d30a-41d5-9d93-bceef0b8938f [ 1028.630248] env[65385]: WARNING neutronclient.v2_0.client [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.631127] env[65385]: WARNING openstack [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.631486] env[65385]: WARNING openstack [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.637112] env[65385]: DEBUG nova.objects.instance [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'pci_requests' on Instance uuid a5fe2ad9-6dfe-485b-bf27-beb218d31467 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.733150] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454249, 'name': ReconfigVM_Task, 'duration_secs': 0.49973} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.733461] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 992920c3-af4f-489e-b21f-9d52c24399de/992920c3-af4f-489e-b21f-9d52c24399de.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.734150] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b83d9d8-fc46-4e09-94ce-e5d2d9050156 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.741525] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1028.741525] env[65385]: value = "task-4454251" [ 1028.741525] env[65385]: _type = "Task" [ 1028.741525] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.758923] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454251, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.813292] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454250, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.820724] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188be83f-76cd-4f97-89d8-af724bb063af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.830849] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09369734-69a9-4595-9be7-f8ef1b920b68 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.871115] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d830047e-c448-419b-a29d-1dfc1e544f47 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.881172] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cbbb65-5d79-4a21-a26b-3ebc820d0f98 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.897912] env[65385]: DEBUG nova.compute.provider_tree [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.054892] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be7974fc-dd04-4b1e-bc20-354f9a8fdb1a tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 41.566s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.147041] env[65385]: DEBUG nova.objects.base [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1029.147041] env[65385]: DEBUG nova.network.neutron [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1029.147041] env[65385]: WARNING neutronclient.v2_0.client [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.147041] env[65385]: WARNING neutronclient.v2_0.client [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.147041] env[65385]: WARNING openstack [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.147041] env[65385]: WARNING openstack [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.258536] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454251, 'name': Rename_Task, 'duration_secs': 0.172505} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.258536] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.258536] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e0cdb42-032b-4c8f-9374-996585bf1774 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.265556] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1029.265556] env[65385]: value = "task-4454252" [ 1029.265556] env[65385]: _type = "Task" [ 1029.265556] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.282614] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1db7f25d-bf16-4f0e-8966-2b15e9299010 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.157s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.284044] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.312712] env[65385]: DEBUG oslo_vmware.api [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454250, 'name': PowerOnVM_Task, 'duration_secs': 0.705701} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.313336] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1029.314210] env[65385]: INFO nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Took 9.96 seconds to spawn the instance on the hypervisor. [ 1029.314210] env[65385]: DEBUG nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1029.315116] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61d50e6-2d8d-4fc0-9b6f-aef795207f24 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.401789] env[65385]: DEBUG nova.scheduler.client.report [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1029.777605] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.838085] env[65385]: INFO nova.compute.manager [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Took 21.62 seconds to build instance. [ 1029.911034] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.375s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.911034] env[65385]: INFO nova.compute.manager [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Migrating [ 1029.917971] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.443s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.271852] env[65385]: INFO nova.compute.manager [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Rebuilding instance [ 1030.281369] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.338422] env[65385]: DEBUG nova.compute.manager [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1030.339383] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbf1bc5-80f2-4214-9f5b-60a33e4a8901 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.344035] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8998f035-150c-4d01-b2fe-287eb9abf108 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.132s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.435910] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.435910] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.436109] env[65385]: DEBUG nova.network.neutron [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1030.440200] env[65385]: INFO nova.compute.claims [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.778102] env[65385]: DEBUG oslo_vmware.api [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454252, 'name': PowerOnVM_Task, 'duration_secs': 1.417049} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.778396] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.778597] env[65385]: INFO nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Took 9.10 seconds to spawn the instance on the hypervisor. [ 1030.778856] env[65385]: DEBUG nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1030.779795] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52874f4-1126-46be-a836-d47bf44abec8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.895801] env[65385]: DEBUG nova.compute.manager [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Received event network-changed-2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1030.896616] env[65385]: DEBUG nova.compute.manager [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Refreshing instance network info cache due to event network-changed-2ea48afc-dea7-435d-996b-e1f73fcec4bf. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1030.896616] env[65385]: DEBUG oslo_concurrency.lockutils [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Acquiring lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.896616] env[65385]: DEBUG oslo_concurrency.lockutils [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Acquired lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.896783] env[65385]: DEBUG nova.network.neutron [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Refreshing network info cache for port 2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1030.945757] env[65385]: WARNING neutronclient.v2_0.client [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.946957] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.947722] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.956591] env[65385]: INFO nova.compute.resource_tracker [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating resource usage from migration e3a9f52b-e795-4f9a-92e7-8b80ef42a93d [ 1031.068095] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.068732] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.154449] env[65385]: WARNING neutronclient.v2_0.client [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.155167] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.155523] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.167971] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ef1e53-fcc0-4641-abd6-ef388f520fbe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.180382] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d30374e-a909-4abd-adb9-d92ab5e74179 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.218345] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62061c2a-96b9-4629-85b2-e831dad71313 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.228352] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263dc72a-b539-47e0-adf7-a42eba6d4120 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.244532] env[65385]: DEBUG nova.compute.provider_tree [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.277523] env[65385]: DEBUG nova.network.neutron [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance_info_cache with network_info: [{"id": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "address": "fa:16:3e:1a:55:36", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26f849c-40", "ovs_interfaceid": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.305473] env[65385]: INFO nova.compute.manager [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Took 22.59 seconds to build instance. [ 1031.360804] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.361319] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e741286c-5ff3-41bd-85ec-0ae91476ff1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.370626] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1031.370626] env[65385]: value = "task-4454253" [ 1031.370626] env[65385]: _type = "Task" [ 1031.370626] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.381693] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.401928] env[65385]: WARNING neutronclient.v2_0.client [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.402471] env[65385]: WARNING openstack [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.402840] env[65385]: WARNING openstack [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.414793] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.415144] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.005s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.415491] env[65385]: DEBUG nova.objects.instance [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'flavor' on Instance uuid a5fe2ad9-6dfe-485b-bf27-beb218d31467 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.580191] env[65385]: WARNING openstack [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.581146] env[65385]: WARNING openstack [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.666338] env[65385]: WARNING neutronclient.v2_0.client [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.666955] env[65385]: WARNING openstack [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.667333] env[65385]: WARNING openstack [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.714692] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.714692] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.748019] env[65385]: DEBUG nova.scheduler.client.report [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1031.773790] env[65385]: DEBUG nova.network.neutron [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updated VIF entry in instance network info cache for port 2ea48afc-dea7-435d-996b-e1f73fcec4bf. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1031.774247] env[65385]: DEBUG nova.network.neutron [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updating instance_info_cache with network_info: [{"id": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "address": "fa:16:3e:00:05:77", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ea48afc-de", "ovs_interfaceid": "2ea48afc-dea7-435d-996b-e1f73fcec4bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.779931] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.813942] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a1bdc0af-9918-436e-8884-05d1d7a717c2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "992920c3-af4f-489e-b21f-9d52c24399de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.108s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.882884] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454253, 'name': PowerOffVM_Task, 'duration_secs': 0.356494} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.883358] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.883716] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.884690] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15896300-dfb9-4b63-9065-9834d84a0aae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.893983] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1031.894423] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d60ef41-c871-4d92-8dfe-0be6e36c5a4f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.923712] env[65385]: WARNING neutronclient.v2_0.client [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.924556] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.924914] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.017947] env[65385]: DEBUG nova.objects.instance [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'pci_requests' on Instance uuid a5fe2ad9-6dfe-485b-bf27-beb218d31467 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.188970] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1032.188970] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1032.188970] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleting the datastore file [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.189304] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e91b6e90-510b-44cf-9bb4-f848d42bc607 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.198145] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1032.198145] env[65385]: value = "task-4454258" [ 1032.198145] env[65385]: _type = "Task" [ 1032.198145] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.206731] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.220799] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.220799] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.220799] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.221161] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.221161] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.221252] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1032.253614] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.335s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.253721] env[65385]: INFO nova.compute.manager [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Migrating [ 1032.278857] env[65385]: DEBUG oslo_concurrency.lockutils [req-4e4671be-d55e-439e-9b3f-94a6c852243a req-da5fe6ba-c324-46be-9214-89d0a918da30 service nova] Releasing lock "refresh_cache-5470afe7-6dd0-4778-9b83-2956ede04b7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.522140] env[65385]: DEBUG nova.objects.base [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1032.522908] env[65385]: DEBUG nova.network.neutron [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1032.523678] env[65385]: WARNING neutronclient.v2_0.client [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.524802] env[65385]: WARNING neutronclient.v2_0.client [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.525891] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.526379] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.577451] env[65385]: DEBUG nova.policy [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1032.711022] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.488136} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.711224] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.711542] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1032.711795] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1032.771602] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.772059] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.772059] env[65385]: DEBUG nova.network.neutron [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1032.931616] env[65385]: DEBUG nova.network.neutron [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Successfully created port: ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1033.278840] env[65385]: WARNING neutronclient.v2_0.client [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1033.279732] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1033.280158] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1033.296130] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d3dcbc-8042-4f0d-a73b-56c2adb4fa03 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.315735] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance '8a974e99-a7b1-438e-aace-9ae82352495c' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1033.523615] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1033.524179] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1033.735035] env[65385]: WARNING neutronclient.v2_0.client [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1033.735814] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1033.736216] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1033.753688] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1033.753987] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1033.754229] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1033.754466] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1033.754645] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1033.754820] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1033.755143] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.755366] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1033.755572] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1033.755757] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1033.755931] env[65385]: DEBUG nova.virt.hardware [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1033.756962] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e99f8f-43f5-4de3-9e9c-78c09810280e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.766693] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e80626-9520-4cd0-8511-a962dd025770 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.782345] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:43:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7e0f7ab-d411-40cb-ae92-d58b922f22fe', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.790188] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1033.790570] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1033.790570] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7abe13fb-0b66-45ed-8bea-d4bff657cb18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.816453] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.816453] env[65385]: value = "task-4454259" [ 1033.816453] env[65385]: _type = "Task" [ 1033.816453] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.821564] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.821834] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc6b3074-c349-43ab-b6aa-022e754e557f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.829536] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454259, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.830925] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1033.830925] env[65385]: value = "task-4454260" [ 1033.830925] env[65385]: _type = "Task" [ 1033.830925] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.842673] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454260, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.865839] env[65385]: DEBUG nova.network.neutron [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1034.229429] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "e13d7fec-250a-412e-b952-f189214c8c1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.229692] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e13d7fec-250a-412e-b952-f189214c8c1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1034.328369] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454259, 'name': CreateVM_Task, 'duration_secs': 0.510997} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.329303] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.329303] env[65385]: WARNING neutronclient.v2_0.client [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1034.329447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.329608] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.330671] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1034.330671] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08b8886f-6854-4e81-bb8a-1058d5573923 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.338866] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1034.338866] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6b1b2-ed5b-77b1-751d-530f0ead9c0a" [ 1034.338866] env[65385]: _type = "Task" [ 1034.338866] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.342436] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454260, 'name': PowerOffVM_Task, 'duration_secs': 0.285449} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.345686] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.345875] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance '8a974e99-a7b1-438e-aace-9ae82352495c' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.355269] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6b1b2-ed5b-77b1-751d-530f0ead9c0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.368484] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.416809] env[65385]: DEBUG nova.compute.manager [req-bf1baed6-9330-49c0-8a1d-c6382d835747 req-219dc64b-0a47-4595-be93-7c6f16b67c63 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-vif-plugged-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1034.416921] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf1baed6-9330-49c0-8a1d-c6382d835747 req-219dc64b-0a47-4595-be93-7c6f16b67c63 service nova] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.417140] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf1baed6-9330-49c0-8a1d-c6382d835747 req-219dc64b-0a47-4595-be93-7c6f16b67c63 service nova] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1034.417305] env[65385]: DEBUG oslo_concurrency.lockutils [req-bf1baed6-9330-49c0-8a1d-c6382d835747 req-219dc64b-0a47-4595-be93-7c6f16b67c63 service nova] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.417466] env[65385]: DEBUG nova.compute.manager [req-bf1baed6-9330-49c0-8a1d-c6382d835747 req-219dc64b-0a47-4595-be93-7c6f16b67c63 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] No waiting events found dispatching network-vif-plugged-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1034.417626] env[65385]: WARNING nova.compute.manager [req-bf1baed6-9330-49c0-8a1d-c6382d835747 req-219dc64b-0a47-4595-be93-7c6f16b67c63 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received unexpected event network-vif-plugged-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 for instance with vm_state active and task_state None. [ 1034.539516] env[65385]: DEBUG nova.network.neutron [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Successfully updated port: ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1034.732605] env[65385]: DEBUG nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1034.855382] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1034.855661] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1034.855817] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1034.855988] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1034.856144] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1034.856285] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1034.856583] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.856755] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1034.856916] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1034.857084] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1034.857252] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1034.863069] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b6b1b2-ed5b-77b1-751d-530f0ead9c0a, 'name': SearchDatastore_Task, 'duration_secs': 0.034135} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.863430] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20c29d29-d442-4263-a34a-8e852703a48a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.874716] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.874958] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.875240] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.875385] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.875567] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.879426] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ce062a9-59f1-489d-afaf-433a3ed66c45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.888835] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1034.888835] env[65385]: value = "task-4454262" [ 1034.888835] env[65385]: _type = "Task" [ 1034.888835] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.898720] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454262, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.900206] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.900871] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.901767] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7744ef8-fe3f-4bb8-b670-ccea2d08ddce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.908292] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1034.908292] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ba73bf-7a45-6bcb-63eb-e2a07c9c0c56" [ 1034.908292] env[65385]: _type = "Task" [ 1034.908292] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.918652] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ba73bf-7a45-6bcb-63eb-e2a07c9c0c56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.042366] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.042621] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1035.042816] env[65385]: DEBUG nova.network.neutron [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1035.259894] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.260295] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.262224] env[65385]: INFO nova.compute.claims [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.400650] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454262, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.419771] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ba73bf-7a45-6bcb-63eb-e2a07c9c0c56, 'name': SearchDatastore_Task, 'duration_secs': 0.0412} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.420698] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8c53339-173d-4c21-800a-660fcd35d0d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.427802] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1035.427802] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529ee24c-ef65-c914-5f81-d0f3fcd2f99d" [ 1035.427802] env[65385]: _type = "Task" [ 1035.427802] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.438404] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529ee24c-ef65-c914-5f81-d0f3fcd2f99d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.546372] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.546907] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.608709] env[65385]: WARNING nova.network.neutron [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] 16967e55-c149-4a58-905f-c47b7facac11 already exists in list: networks containing: ['16967e55-c149-4a58-905f-c47b7facac11']. ignoring it [ 1035.649043] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.649460] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.717227] env[65385]: WARNING neutronclient.v2_0.client [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.717914] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.718319] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.831349] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.831747] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.895682] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370e46da-a297-4c6a-9507-ebadc99e8f89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.899801] env[65385]: WARNING neutronclient.v2_0.client [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.900474] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.900841] env[65385]: WARNING openstack [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.915251] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454262, 'name': ReconfigVM_Task, 'duration_secs': 0.536535} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.931465] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance '8a974e99-a7b1-438e-aace-9ae82352495c' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.935801] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance '59f26752-75ab-45e5-b0a9-bbcef44ecd7f' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.950685] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529ee24c-ef65-c914-5f81-d0f3fcd2f99d, 'name': SearchDatastore_Task, 'duration_secs': 0.050224} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.950949] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1035.951226] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1035.951500] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed6518cc-1f57-4061-b00b-5531825c311d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.961946] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1035.961946] env[65385]: value = "task-4454263" [ 1035.961946] env[65385]: _type = "Task" [ 1035.961946] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.977334] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454263, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.010660] env[65385]: DEBUG nova.network.neutron [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "address": "fa:16:3e:6e:b7:19", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef6b6d8f-dd", "ovs_interfaceid": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1036.442969] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1036.443327] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1036.443385] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1036.443545] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1036.443691] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1036.443836] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1036.444050] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.444305] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1036.444525] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1036.444697] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1036.444872] env[65385]: DEBUG nova.virt.hardware [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1036.450589] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1036.455277] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.455554] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e48480d5-b4a2-4a8a-ae77-6c05e177e216 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.469494] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec8da6e6-c3b8-47b2-b898-56395e5513de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.484718] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454263, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.485169] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1036.485169] env[65385]: value = "task-4454264" [ 1036.485169] env[65385]: _type = "Task" [ 1036.485169] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.487833] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1036.487833] env[65385]: value = "task-4454265" [ 1036.487833] env[65385]: _type = "Task" [ 1036.487833] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.488630] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80af9dc2-c039-4416-9891-09edb655677f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.504828] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.512978] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454265, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.513657] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.514390] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.514581] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1036.515920] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456e0924-3aad-45cf-90c0-a2c4ead3a160 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.520540] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649ba9c4-5a19-40b1-b7f0-cc67ee7db484 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.541388] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1036.541711] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1036.541824] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1036.542056] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1036.542384] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1036.542512] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1036.542691] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.542843] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1036.543054] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1036.543226] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1036.543458] env[65385]: DEBUG nova.virt.hardware [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1036.550482] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfiguring VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1036.581852] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc7a0c52-98f5-4ec5-b67d-1df193dffdda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.597756] env[65385]: DEBUG nova.compute.manager [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-changed-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1036.598057] env[65385]: DEBUG nova.compute.manager [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing instance network info cache due to event network-changed-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1036.598326] env[65385]: DEBUG oslo_concurrency.lockutils [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.598492] env[65385]: DEBUG oslo_concurrency.lockutils [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1036.598660] env[65385]: DEBUG nova.network.neutron [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing network info cache for port ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1036.602139] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ff44ba-4ba3-4642-a3a6-3718bfeda609 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.616977] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ab51f8-b6e7-4251-970a-f5532333edc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.622070] env[65385]: DEBUG oslo_vmware.api [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1036.622070] env[65385]: value = "task-4454266" [ 1036.622070] env[65385]: _type = "Task" [ 1036.622070] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.636280] env[65385]: DEBUG nova.compute.provider_tree [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.640664] env[65385]: DEBUG oslo_vmware.api [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454266, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.983374] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454263, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689907} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.983720] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1036.983954] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.984353] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6ee228f-99ea-4320-bec4-50732ecfa0ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.993297] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1036.993297] env[65385]: value = "task-4454268" [ 1036.993297] env[65385]: _type = "Task" [ 1036.993297] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.999835] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454264, 'name': PowerOffVM_Task, 'duration_secs': 0.354699} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.003854] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.004122] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance '59f26752-75ab-45e5-b0a9-bbcef44ecd7f' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1037.011372] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454268, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.019314] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454265, 'name': ReconfigVM_Task, 'duration_secs': 0.282912} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.019739] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1037.021144] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d448d144-7821-4019-b3b7-f11708926d8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.050963] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 8a974e99-a7b1-438e-aace-9ae82352495c/8a974e99-a7b1-438e-aace-9ae82352495c.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.051356] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5942c9a-502b-4ada-bf74-9b19dcf9b244 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.076476] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1037.076476] env[65385]: value = "task-4454269" [ 1037.076476] env[65385]: _type = "Task" [ 1037.076476] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.089998] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454269, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.102019] env[65385]: WARNING neutronclient.v2_0.client [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.103019] env[65385]: WARNING openstack [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.103489] env[65385]: WARNING openstack [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.134995] env[65385]: DEBUG oslo_vmware.api [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454266, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.142147] env[65385]: DEBUG nova.scheduler.client.report [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1037.214442] env[65385]: WARNING openstack [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.214893] env[65385]: WARNING openstack [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.285060] env[65385]: WARNING neutronclient.v2_0.client [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.285837] env[65385]: WARNING openstack [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.286316] env[65385]: WARNING openstack [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.373571] env[65385]: DEBUG nova.network.neutron [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updated VIF entry in instance network info cache for port ef6b6d8f-dd4a-4682-991b-9c6dd0d27525. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1037.374009] env[65385]: DEBUG nova.network.neutron [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "address": "fa:16:3e:6e:b7:19", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef6b6d8f-dd", "ovs_interfaceid": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1037.505355] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454268, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09231} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.505926] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.506987] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a3852d-a2f7-4bad-accc-facdbca336f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.515031] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1037.515301] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1037.515476] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1037.515799] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1037.516040] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1037.516226] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1037.516454] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.516633] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1037.516803] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1037.516969] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1037.517187] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1037.536191] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b879e347-ed06-4798-a059-b77d70bd31c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.555809] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.556191] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aff678fa-7d09-44ed-959d-60239bed0e89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.579548] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1037.579548] env[65385]: value = "task-4454270" [ 1037.579548] env[65385]: _type = "Task" [ 1037.579548] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.584135] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1037.584135] env[65385]: value = "task-4454271" [ 1037.584135] env[65385]: _type = "Task" [ 1037.584135] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.595027] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.601976] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454271, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.602358] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454270, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.634797] env[65385]: DEBUG oslo_vmware.api [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454266, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.647539] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.648138] env[65385]: DEBUG nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1037.877733] env[65385]: DEBUG oslo_concurrency.lockutils [req-f61ef7e2-202f-4ff1-a135-945f78646ef3 req-a2ff4436-2a67-48c5-85be-77c2035670c0 service nova] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.096795] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454269, 'name': ReconfigVM_Task, 'duration_secs': 0.854434} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.100383] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 8a974e99-a7b1-438e-aace-9ae82352495c/8a974e99-a7b1-438e-aace-9ae82352495c.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.100635] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance '8a974e99-a7b1-438e-aace-9ae82352495c' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1038.104204] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454270, 'name': ReconfigVM_Task, 'duration_secs': 0.347006} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.107944] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance '59f26752-75ab-45e5-b0a9-bbcef44ecd7f' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1038.111371] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454271, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.135134] env[65385]: DEBUG oslo_vmware.api [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454266, 'name': ReconfigVM_Task, 'duration_secs': 1.266676} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.135631] env[65385]: WARNING neutronclient.v2_0.client [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.135877] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.136100] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfigured VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1038.152632] env[65385]: DEBUG nova.compute.utils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1038.155211] env[65385]: DEBUG nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1038.155448] env[65385]: DEBUG nova.network.neutron [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1038.155760] env[65385]: WARNING neutronclient.v2_0.client [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.156064] env[65385]: WARNING neutronclient.v2_0.client [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.156651] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1038.157021] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.210605] env[65385]: DEBUG nova.policy [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662871f375e441eb98d814cf7d7ac5c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bc8afc699e34f059f18d8244980bc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1038.516593] env[65385]: DEBUG nova.network.neutron [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Successfully created port: 3f844c95-3619-44bb-945c-bec15b1f30ec {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1038.600222] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454271, 'name': ReconfigVM_Task, 'duration_secs': 0.54865} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.600501] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d/537393de-ea48-4d34-8e04-95331bbc099d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.601148] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d32609ee-deb1-4f3b-833c-0e69ed1e5120 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.609026] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1038.609026] env[65385]: value = "task-4454272" [ 1038.609026] env[65385]: _type = "Task" [ 1038.609026] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.616668] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1038.616933] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1038.617048] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1038.617382] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1038.617486] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1038.617630] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1038.617856] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.618047] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1038.618233] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1038.618401] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1038.618570] env[65385]: DEBUG nova.virt.hardware [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1038.624050] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Reconfiguring VM instance instance-0000001e to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1038.625060] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b36691-e40c-4cc9-b7fc-ce62fff65d3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.627942] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7058bce7-c52b-445c-a1bb-e4ddc286ad11 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.643809] env[65385]: DEBUG oslo_concurrency.lockutils [None req-42811cf2-b103-4514-afea-ff852564b477 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.229s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.649266] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454272, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.672085] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1038.672085] env[65385]: value = "task-4454273" [ 1038.672085] env[65385]: _type = "Task" [ 1038.672085] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.672085] env[65385]: DEBUG nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1038.678164] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a557d5-4f4a-4848-bcde-06b8bc309f37 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.692018] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.710937] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance '8a974e99-a7b1-438e-aace-9ae82352495c' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1039.121924] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454272, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.190793] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454273, 'name': ReconfigVM_Task, 'duration_secs': 0.183035} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.191210] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Reconfigured VM instance instance-0000001e to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1039.192231] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c9dd79-4f3d-4b13-a784-f2db1b7f9329 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.215737] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 59f26752-75ab-45e5-b0a9-bbcef44ecd7f/59f26752-75ab-45e5-b0a9-bbcef44ecd7f.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.218306] env[65385]: WARNING neutronclient.v2_0.client [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.218636] env[65385]: WARNING neutronclient.v2_0.client [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.220748] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f01f5a58-5e9b-40cd-9f5b-4ecc1335edc7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.240669] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1039.240669] env[65385]: value = "task-4454275" [ 1039.240669] env[65385]: _type = "Task" [ 1039.240669] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.250299] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.268952] env[65385]: DEBUG nova.network.neutron [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Port f26f849c-4075-4ee1-b174-88ce8e66b69c binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1039.620481] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454272, 'name': Rename_Task, 'duration_secs': 0.655263} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.620860] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1039.620992] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24d41451-c661-4541-bfa9-feb55144988c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.628531] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1039.628531] env[65385]: value = "task-4454276" [ 1039.628531] env[65385]: _type = "Task" [ 1039.628531] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.637982] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454276, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.686966] env[65385]: DEBUG nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1039.715095] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1039.715393] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1039.715563] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1039.715744] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1039.715884] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1039.716034] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1039.716241] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1039.716402] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1039.716566] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1039.716723] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1039.716920] env[65385]: DEBUG nova.virt.hardware [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1039.717750] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3327b45-d5ee-47a1-a2d0-14366364af3f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.725845] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d012ee26-aeed-4532-9a82-763e35fd7878 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.749959] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454275, 'name': ReconfigVM_Task, 'duration_secs': 0.304507} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.750269] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 59f26752-75ab-45e5-b0a9-bbcef44ecd7f/59f26752-75ab-45e5-b0a9-bbcef44ecd7f.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.750540] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance '59f26752-75ab-45e5-b0a9-bbcef44ecd7f' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1040.035078] env[65385]: DEBUG nova.compute.manager [req-0e038dc5-95cf-46ef-8007-eca3a8453756 req-3b0c7bbd-3759-4e6f-b0e8-7d8b4972a864 service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Received event network-vif-plugged-3f844c95-3619-44bb-945c-bec15b1f30ec {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1040.035078] env[65385]: DEBUG oslo_concurrency.lockutils [req-0e038dc5-95cf-46ef-8007-eca3a8453756 req-3b0c7bbd-3759-4e6f-b0e8-7d8b4972a864 service nova] Acquiring lock "e13d7fec-250a-412e-b952-f189214c8c1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.035078] env[65385]: DEBUG oslo_concurrency.lockutils [req-0e038dc5-95cf-46ef-8007-eca3a8453756 req-3b0c7bbd-3759-4e6f-b0e8-7d8b4972a864 service nova] Lock "e13d7fec-250a-412e-b952-f189214c8c1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.035078] env[65385]: DEBUG oslo_concurrency.lockutils [req-0e038dc5-95cf-46ef-8007-eca3a8453756 req-3b0c7bbd-3759-4e6f-b0e8-7d8b4972a864 service nova] Lock "e13d7fec-250a-412e-b952-f189214c8c1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.035078] env[65385]: DEBUG nova.compute.manager [req-0e038dc5-95cf-46ef-8007-eca3a8453756 req-3b0c7bbd-3759-4e6f-b0e8-7d8b4972a864 service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] No waiting events found dispatching network-vif-plugged-3f844c95-3619-44bb-945c-bec15b1f30ec {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1040.035078] env[65385]: WARNING nova.compute.manager [req-0e038dc5-95cf-46ef-8007-eca3a8453756 req-3b0c7bbd-3759-4e6f-b0e8-7d8b4972a864 service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Received unexpected event network-vif-plugged-3f844c95-3619-44bb-945c-bec15b1f30ec for instance with vm_state building and task_state spawning. [ 1040.077047] env[65385]: DEBUG nova.network.neutron [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Successfully updated port: 3f844c95-3619-44bb-945c-bec15b1f30ec {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1040.142173] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454276, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.257441] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0153bb5-ab4e-494b-a788-5e982778aea4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.287303] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ccc7bd-7365-41f3-a839-f8e353d00e7e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.297681] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.297933] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.298116] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.317091] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance '59f26752-75ab-45e5-b0a9-bbcef44ecd7f' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1040.549055] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.549339] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.549694] env[65385]: DEBUG nova.objects.instance [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'flavor' on Instance uuid a5fe2ad9-6dfe-485b-bf27-beb218d31467 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.580319] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "refresh_cache-e13d7fec-250a-412e-b952-f189214c8c1e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.580462] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "refresh_cache-e13d7fec-250a-412e-b952-f189214c8c1e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1040.580611] env[65385]: DEBUG nova.network.neutron [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1040.640342] env[65385]: DEBUG oslo_vmware.api [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454276, 'name': PowerOnVM_Task, 'duration_secs': 0.811018} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.640809] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.640809] env[65385]: DEBUG nova.compute.manager [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1040.641661] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a8ddaf-676f-4069-b658-df06d3d9ae44 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.824358] env[65385]: WARNING neutronclient.v2_0.client [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.824764] env[65385]: WARNING neutronclient.v2_0.client [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.866470] env[65385]: DEBUG nova.network.neutron [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Port 59289962-8634-414f-9bd2-b3f5b19af9b4 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1041.053471] env[65385]: WARNING neutronclient.v2_0.client [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.054709] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.055187] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.083511] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.083973] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.126394] env[65385]: DEBUG nova.network.neutron [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1041.149948] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.150395] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.165039] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1041.165300] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.165528] env[65385]: DEBUG nova.objects.instance [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1041.195527] env[65385]: DEBUG nova.objects.instance [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'pci_requests' on Instance uuid a5fe2ad9-6dfe-485b-bf27-beb218d31467 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.228547] env[65385]: WARNING neutronclient.v2_0.client [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.229580] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.229927] env[65385]: WARNING openstack [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.305372] env[65385]: WARNING neutronclient.v2_0.client [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.321909] env[65385]: DEBUG nova.network.neutron [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Updating instance_info_cache with network_info: [{"id": "3f844c95-3619-44bb-945c-bec15b1f30ec", "address": "fa:16:3e:d4:c6:5f", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f844c95-36", "ovs_interfaceid": "3f844c95-3619-44bb-945c-bec15b1f30ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1041.351505] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.351688] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1041.351857] env[65385]: DEBUG nova.network.neutron [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1041.501909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "537393de-ea48-4d34-8e04-95331bbc099d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1041.502392] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "537393de-ea48-4d34-8e04-95331bbc099d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.502723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "537393de-ea48-4d34-8e04-95331bbc099d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1041.503011] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "537393de-ea48-4d34-8e04-95331bbc099d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.503250] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "537393de-ea48-4d34-8e04-95331bbc099d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.506149] env[65385]: INFO nova.compute.manager [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Terminating instance [ 1041.699189] env[65385]: DEBUG nova.objects.base [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1041.699587] env[65385]: DEBUG nova.network.neutron [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1041.699844] env[65385]: WARNING neutronclient.v2_0.client [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.700194] env[65385]: WARNING neutronclient.v2_0.client [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.700836] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.701284] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.782405] env[65385]: DEBUG nova.policy [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1041.824835] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "refresh_cache-e13d7fec-250a-412e-b952-f189214c8c1e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1041.825172] env[65385]: DEBUG nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Instance network_info: |[{"id": "3f844c95-3619-44bb-945c-bec15b1f30ec", "address": "fa:16:3e:d4:c6:5f", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f844c95-36", "ovs_interfaceid": "3f844c95-3619-44bb-945c-bec15b1f30ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1041.825672] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:c6:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f844c95-3619-44bb-945c-bec15b1f30ec', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1041.833527] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1041.833782] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1041.834545] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9284bc6-474f-4b82-8a03-0f5e22ccdac7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.854386] env[65385]: WARNING neutronclient.v2_0.client [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.855093] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.855518] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.863905] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1041.863905] env[65385]: value = "task-4454278" [ 1041.863905] env[65385]: _type = "Task" [ 1041.863905] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.873441] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454278, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.889656] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1041.889656] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.889904] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.963176] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.963581] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.010776] env[65385]: DEBUG nova.compute.manager [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1042.011049] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.012050] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fa2d98-e001-45bc-a9e1-9889d1df962f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.023665] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.024033] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4a0ffe1-5903-479e-b432-afb0b6c68e18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.032473] env[65385]: DEBUG oslo_vmware.api [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1042.032473] env[65385]: value = "task-4454279" [ 1042.032473] env[65385]: _type = "Task" [ 1042.032473] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.037597] env[65385]: WARNING neutronclient.v2_0.client [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.038302] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.038685] env[65385]: WARNING openstack [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.059905] env[65385]: DEBUG oslo_vmware.api [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.076109] env[65385]: DEBUG nova.compute.manager [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Received event network-changed-3f844c95-3619-44bb-945c-bec15b1f30ec {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1042.076329] env[65385]: DEBUG nova.compute.manager [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Refreshing instance network info cache due to event network-changed-3f844c95-3619-44bb-945c-bec15b1f30ec. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1042.076570] env[65385]: DEBUG oslo_concurrency.lockutils [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Acquiring lock "refresh_cache-e13d7fec-250a-412e-b952-f189214c8c1e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.076718] env[65385]: DEBUG oslo_concurrency.lockutils [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Acquired lock "refresh_cache-e13d7fec-250a-412e-b952-f189214c8c1e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.076875] env[65385]: DEBUG nova.network.neutron [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Refreshing network info cache for port 3f844c95-3619-44bb-945c-bec15b1f30ec {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1042.086746] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "77d718c8-4959-483a-8717-c00f2be1d0bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.087062] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.138102] env[65385]: DEBUG nova.network.neutron [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance_info_cache with network_info: [{"id": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "address": "fa:16:3e:1a:55:36", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26f849c-40", "ovs_interfaceid": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1042.175504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-76fa9850-b347-4a42-8085-eff69c7ebfa9 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.374469] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454278, 'name': CreateVM_Task, 'duration_secs': 0.421164} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.374678] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1042.376042] env[65385]: WARNING neutronclient.v2_0.client [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.376042] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.376042] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.376042] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1042.376315] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f182d2da-6539-4689-ac31-ae0f1f764219 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.381138] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1042.381138] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528beaa0-e184-65d6-379f-3b969a789f5d" [ 1042.381138] env[65385]: _type = "Task" [ 1042.381138] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.389225] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528beaa0-e184-65d6-379f-3b969a789f5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.542663] env[65385]: DEBUG oslo_vmware.api [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454279, 'name': PowerOffVM_Task, 'duration_secs': 0.237989} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.542942] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1042.543122] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1042.543386] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2362757-3bdb-4821-b440-5e1804471672 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.580462] env[65385]: WARNING neutronclient.v2_0.client [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.581245] env[65385]: WARNING openstack [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.581637] env[65385]: WARNING openstack [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.590238] env[65385]: DEBUG nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1042.641417] env[65385]: DEBUG oslo_concurrency.lockutils [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1042.695063] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1042.695063] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1042.695289] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleting the datastore file [datastore1] 537393de-ea48-4d34-8e04-95331bbc099d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.695719] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b9e7195-e0b8-44f5-a092-899db2ce8c8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.701751] env[65385]: WARNING openstack [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.702300] env[65385]: WARNING openstack [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.711617] env[65385]: DEBUG oslo_vmware.api [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1042.711617] env[65385]: value = "task-4454281" [ 1042.711617] env[65385]: _type = "Task" [ 1042.711617] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.722761] env[65385]: DEBUG oslo_vmware.api [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454281, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.852415] env[65385]: WARNING neutronclient.v2_0.client [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.853402] env[65385]: WARNING openstack [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.853868] env[65385]: WARNING openstack [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.893040] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528beaa0-e184-65d6-379f-3b969a789f5d, 'name': SearchDatastore_Task, 'duration_secs': 0.009781} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.893565] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1042.893989] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1042.894400] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.894640] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.894960] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1042.896859] env[65385]: WARNING neutronclient.v2_0.client [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.900020] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a99038ba-3e9d-45c1-a7a8-7bc44ec5ad53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.914673] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1042.914871] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1042.915683] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc5cce0b-8610-4db9-b9eb-db5aee96b34e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.922464] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1042.922464] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b33927-6a6b-983d-7270-c969bb37a06f" [ 1042.922464] env[65385]: _type = "Task" [ 1042.922464] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.932495] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b33927-6a6b-983d-7270-c969bb37a06f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.961445] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.961634] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.961806] env[65385]: DEBUG nova.network.neutron [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1042.994111] env[65385]: DEBUG nova.network.neutron [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Updated VIF entry in instance network info cache for port 3f844c95-3619-44bb-945c-bec15b1f30ec. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1042.994505] env[65385]: DEBUG nova.network.neutron [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Updating instance_info_cache with network_info: [{"id": "3f844c95-3619-44bb-945c-bec15b1f30ec", "address": "fa:16:3e:d4:c6:5f", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f844c95-36", "ovs_interfaceid": "3f844c95-3619-44bb-945c-bec15b1f30ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.117016] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1043.117306] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1043.118873] env[65385]: INFO nova.compute.claims [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.163106] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868fef17-c3b6-4f39-84a3-87e19f68e285 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.183138] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4599ab6a-7aa9-4215-91ff-0ea2e91ade0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.190815] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance '8a974e99-a7b1-438e-aace-9ae82352495c' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1043.222745] env[65385]: DEBUG oslo_vmware.api [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454281, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197157} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.223139] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.223300] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.223348] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1043.223511] env[65385]: INFO nova.compute.manager [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1043.223781] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1043.224088] env[65385]: DEBUG nova.compute.manager [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1043.224088] env[65385]: DEBUG nova.network.neutron [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1043.224418] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.225097] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.225353] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.262661] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.328103] env[65385]: DEBUG nova.network.neutron [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Successfully updated port: 05d2fd35-d0a1-44d7-9944-0b6445e2bd9b {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1043.436298] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b33927-6a6b-983d-7270-c969bb37a06f, 'name': SearchDatastore_Task, 'duration_secs': 0.010442} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.437198] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4b081fc-7121-4d1d-9a65-a75bfe8ebd5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.443752] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1043.443752] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52592c73-2d27-707b-23dc-9a69ad5a8480" [ 1043.443752] env[65385]: _type = "Task" [ 1043.443752] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.453424] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52592c73-2d27-707b-23dc-9a69ad5a8480, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.465617] env[65385]: WARNING neutronclient.v2_0.client [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.466441] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.466842] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.498238] env[65385]: DEBUG oslo_concurrency.lockutils [req-a21cef3d-5cf5-4e51-9d08-a8b4146cb5d6 req-d9d5ff0e-aff8-434f-96d4-dc9fcc0f970d service nova] Releasing lock "refresh_cache-e13d7fec-250a-412e-b952-f189214c8c1e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.509235] env[65385]: DEBUG nova.compute.manager [req-8a0cc7aa-8130-4d37-bd7a-9a4f52824721 req-4f39b6db-2d4d-4bcf-88ce-7f42609d20ce service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Received event network-vif-deleted-a7e0f7ab-d411-40cb-ae92-d58b922f22fe {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1043.509445] env[65385]: INFO nova.compute.manager [req-8a0cc7aa-8130-4d37-bd7a-9a4f52824721 req-4f39b6db-2d4d-4bcf-88ce-7f42609d20ce service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Neutron deleted interface a7e0f7ab-d411-40cb-ae92-d58b922f22fe; detaching it from the instance and deleting it from the info cache [ 1043.509635] env[65385]: DEBUG nova.network.neutron [req-8a0cc7aa-8130-4d37-bd7a-9a4f52824721 req-4f39b6db-2d4d-4bcf-88ce-7f42609d20ce service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.619675] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.620095] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.686731] env[65385]: WARNING neutronclient.v2_0.client [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.687382] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.687728] env[65385]: WARNING openstack [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.696976] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1043.697335] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc72a166-7469-40ac-af27-fa9bb63d9c30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.705702] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1043.705702] env[65385]: value = "task-4454283" [ 1043.705702] env[65385]: _type = "Task" [ 1043.705702] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.722087] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.773773] env[65385]: DEBUG nova.network.neutron [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.828654] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.828896] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.829221] env[65385]: DEBUG nova.network.neutron [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1043.956711] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52592c73-2d27-707b-23dc-9a69ad5a8480, 'name': SearchDatastore_Task, 'duration_secs': 0.020799} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.957040] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.957307] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e13d7fec-250a-412e-b952-f189214c8c1e/e13d7fec-250a-412e-b952-f189214c8c1e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1043.957608] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-343bead9-5c40-4a90-a93b-c1beb235c03d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.965737] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1043.965737] env[65385]: value = "task-4454284" [ 1043.965737] env[65385]: _type = "Task" [ 1043.965737] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.974787] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454284, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.984310] env[65385]: DEBUG nova.network.neutron [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1044.012760] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e0809d7-fa9f-4ef5-a200-161b8f51893d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.024264] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20dec5c5-5d8b-49dd-a76a-08b8dd324f6c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.059852] env[65385]: DEBUG nova.compute.manager [req-8a0cc7aa-8130-4d37-bd7a-9a4f52824721 req-4f39b6db-2d4d-4bcf-88ce-7f42609d20ce service nova] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Detach interface failed, port_id=a7e0f7ab-d411-40cb-ae92-d58b922f22fe, reason: Instance 537393de-ea48-4d34-8e04-95331bbc099d could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1044.206975] env[65385]: DEBUG nova.compute.manager [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-vif-plugged-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1044.207339] env[65385]: DEBUG oslo_concurrency.lockutils [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.207413] env[65385]: DEBUG oslo_concurrency.lockutils [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.207611] env[65385]: DEBUG oslo_concurrency.lockutils [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.207799] env[65385]: DEBUG nova.compute.manager [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] No waiting events found dispatching network-vif-plugged-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1044.207985] env[65385]: WARNING nova.compute.manager [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received unexpected event network-vif-plugged-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b for instance with vm_state active and task_state None. [ 1044.208133] env[65385]: DEBUG nova.compute.manager [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-changed-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1044.208254] env[65385]: DEBUG nova.compute.manager [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing instance network info cache due to event network-changed-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1044.208435] env[65385]: DEBUG oslo_concurrency.lockutils [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.222542] env[65385]: DEBUG oslo_vmware.api [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454283, 'name': PowerOnVM_Task, 'duration_secs': 0.483938} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.222929] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1044.223130] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-39838505-be08-44ca-888b-eb6e2c7bbb22 tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance '8a974e99-a7b1-438e-aace-9ae82352495c' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1044.276581] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.332654] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.333105] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.389489] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c575fae9-9a9b-46f3-881b-a0d3d054185a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.393963] env[65385]: WARNING nova.network.neutron [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] 16967e55-c149-4a58-905f-c47b7facac11 already exists in list: networks containing: ['16967e55-c149-4a58-905f-c47b7facac11']. ignoring it [ 1044.394168] env[65385]: WARNING nova.network.neutron [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] 16967e55-c149-4a58-905f-c47b7facac11 already exists in list: networks containing: ['16967e55-c149-4a58-905f-c47b7facac11']. ignoring it [ 1044.403478] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ed975f-10a3-4218-839e-9df4c20e2ba1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.441781] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.442202] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.449235] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb3516fb-392e-4d95-8656-8c7918a5fa1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.458279] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41dd86c-1dae-4c68-b32e-75c3c8a7ec3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.473131] env[65385]: DEBUG nova.compute.provider_tree [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.489283] env[65385]: INFO nova.compute.manager [-] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Took 1.27 seconds to deallocate network for instance. [ 1044.489863] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454284, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481637} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.491840] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e13d7fec-250a-412e-b952-f189214c8c1e/e13d7fec-250a-412e-b952-f189214c8c1e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1044.491840] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.492011] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-923cf07b-1fe0-482d-88d5-4b3564dd33c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.503881] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1044.503881] env[65385]: value = "task-4454285" [ 1044.503881] env[65385]: _type = "Task" [ 1044.503881] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.514295] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454285, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.516061] env[65385]: WARNING neutronclient.v2_0.client [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1044.516709] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.517080] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.644051] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.644804] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.705734] env[65385]: WARNING neutronclient.v2_0.client [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1044.706416] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.706768] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.805110] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf21e1f9-02a9-46d3-b381-1a3146c6bee3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.825322] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.825706] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.832447] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-838075c7-5532-4493-a652-241a964756f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.841599] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance '59f26752-75ab-45e5-b0a9-bbcef44ecd7f' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1044.893063] env[65385]: WARNING neutronclient.v2_0.client [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1044.893422] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.894172] env[65385]: WARNING openstack [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.976195] env[65385]: DEBUG nova.network.neutron [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "address": "fa:16:3e:6e:b7:19", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef6b6d8f-dd", "ovs_interfaceid": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "05d2fd35-d0a1-44d7-9944-0b6445e2bd9b", "address": "fa:16:3e:2f:de:4d", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05d2fd35-d0", "ovs_interfaceid": "05d2fd35-d0a1-44d7-9944-0b6445e2bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1044.978765] env[65385]: DEBUG nova.scheduler.client.report [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1044.999217] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.015085] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454285, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067197} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.015085] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.015845] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873b7c65-86eb-4fd2-85c4-59b0bd6d421c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.039648] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] e13d7fec-250a-412e-b952-f189214c8c1e/e13d7fec-250a-412e-b952-f189214c8c1e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.040025] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5142c1cf-90e2-49ee-b8ec-64c4db0f55a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.061913] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1045.061913] env[65385]: value = "task-4454286" [ 1045.061913] env[65385]: _type = "Task" [ 1045.061913] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.072603] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454286, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.348143] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1045.348612] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a34cf7ba-2a6b-4f28-b323-1392fe6368f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.357905] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1045.357905] env[65385]: value = "task-4454288" [ 1045.357905] env[65385]: _type = "Task" [ 1045.357905] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.370672] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454288, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.484355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1045.485328] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.485541] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.486422] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.487033] env[65385]: DEBUG nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1045.489524] env[65385]: DEBUG oslo_concurrency.lockutils [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.489789] env[65385]: DEBUG nova.network.neutron [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Refreshing network info cache for port 05d2fd35-d0a1-44d7-9944-0b6445e2bd9b {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1045.492543] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac41ace7-cf7a-4089-9761-e84323e1736a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.495548] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.497s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.495805] env[65385]: DEBUG nova.objects.instance [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lazy-loading 'resources' on Instance uuid 537393de-ea48-4d34-8e04-95331bbc099d {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.516719] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1045.517053] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1045.517288] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1045.517547] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1045.517769] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1045.517983] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1045.518268] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.518477] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1045.518723] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1045.518964] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1045.519234] env[65385]: DEBUG nova.virt.hardware [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1045.527016] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfiguring VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1045.529103] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d49d0a22-1076-4853-ad7a-c21a9200bef4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.548969] env[65385]: DEBUG oslo_vmware.api [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1045.548969] env[65385]: value = "task-4454289" [ 1045.548969] env[65385]: _type = "Task" [ 1045.548969] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.558495] env[65385]: DEBUG oslo_vmware.api [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454289, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.575371] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454286, 'name': ReconfigVM_Task, 'duration_secs': 0.312951} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.577057] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Reconfigured VM instance instance-00000060 to attach disk [datastore2] e13d7fec-250a-412e-b952-f189214c8c1e/e13d7fec-250a-412e-b952-f189214c8c1e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.577057] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05599638-5fa2-485a-b5d9-969110a8b797 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.583517] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1045.583517] env[65385]: value = "task-4454290" [ 1045.583517] env[65385]: _type = "Task" [ 1045.583517] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.592423] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454290, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.868833] env[65385]: DEBUG oslo_vmware.api [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454288, 'name': PowerOnVM_Task, 'duration_secs': 0.462286} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.869359] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.869450] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98d0d63b-c3c1-400f-83e2-c0b614aa9e16 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance '59f26752-75ab-45e5-b0a9-bbcef44ecd7f' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1045.999043] env[65385]: WARNING neutronclient.v2_0.client [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.999759] env[65385]: WARNING openstack [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.000128] env[65385]: WARNING openstack [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.009440] env[65385]: DEBUG nova.compute.utils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1046.014450] env[65385]: DEBUG nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1046.014670] env[65385]: DEBUG nova.network.neutron [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1046.014899] env[65385]: WARNING neutronclient.v2_0.client [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.015244] env[65385]: WARNING neutronclient.v2_0.client [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.015820] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.016169] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.024844] env[65385]: DEBUG nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1046.064723] env[65385]: DEBUG oslo_vmware.api [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454289, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.084381] env[65385]: DEBUG nova.policy [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c1a9d4194964403a8d11abfad4c65a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94cc5bc221c0455ea760f5022db6bdc4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1046.107025] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454290, 'name': Rename_Task, 'duration_secs': 0.22609} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.107288] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.107558] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d819dffd-576f-4562-adf7-af2da55e33c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.117362] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1046.117362] env[65385]: value = "task-4454291" [ 1046.117362] env[65385]: _type = "Task" [ 1046.117362] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.130013] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454291, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.180186] env[65385]: WARNING openstack [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.180853] env[65385]: WARNING openstack [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.262218] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d973318-a125-4b68-9128-e4057a666705 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.266308] env[65385]: WARNING neutronclient.v2_0.client [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.267125] env[65385]: WARNING openstack [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.267506] env[65385]: WARNING openstack [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.285839] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc8e968-0eb3-4cc1-bc68-0d318f4be4b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.329905] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b616cc6-8e64-4ba0-926a-a661aba967d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.343438] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4ab50e-745f-4a66-8297-66728cf6d127 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.361562] env[65385]: DEBUG nova.compute.provider_tree [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.389214] env[65385]: DEBUG nova.network.neutron [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updated VIF entry in instance network info cache for port 05d2fd35-d0a1-44d7-9944-0b6445e2bd9b. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1046.389318] env[65385]: DEBUG nova.network.neutron [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "address": "fa:16:3e:6e:b7:19", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef6b6d8f-dd", "ovs_interfaceid": "ef6b6d8f-dd4a-4682-991b-9c6dd0d27525", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "05d2fd35-d0a1-44d7-9944-0b6445e2bd9b", "address": "fa:16:3e:2f:de:4d", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05d2fd35-d0", "ovs_interfaceid": "05d2fd35-d0a1-44d7-9944-0b6445e2bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1046.411430] env[65385]: DEBUG nova.network.neutron [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Successfully created port: 0fc23e28-1265-46b7-aa43-fd72e2269330 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1046.564650] env[65385]: DEBUG oslo_vmware.api [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454289, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.634520] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454291, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.839156] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "8a974e99-a7b1-438e-aace-9ae82352495c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.839429] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.839508] env[65385]: DEBUG nova.compute.manager [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Going to confirm migration 4 {{(pid=65385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1046.865237] env[65385]: DEBUG nova.scheduler.client.report [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1046.891954] env[65385]: DEBUG oslo_concurrency.lockutils [req-2091f150-4c32-4272-8986-01f8fc60dbfb req-2d2a48a5-5f31-45cd-ae15-7e058c5d1ee5 service nova] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.035536] env[65385]: DEBUG nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1047.065405] env[65385]: DEBUG oslo_vmware.api [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454289, 'name': ReconfigVM_Task, 'duration_secs': 1.097447} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.067855] env[65385]: WARNING neutronclient.v2_0.client [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.068167] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.068381] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfigured VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1047.078641] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1047.078926] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1047.079471] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1047.079471] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1047.079471] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1047.079703] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1047.079834] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1047.079960] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1047.080099] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1047.080260] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1047.080449] env[65385]: DEBUG nova.virt.hardware [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1047.081411] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf27713b-15f7-4139-9116-a41acccabf4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.091067] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7e02dd-0347-43dd-b896-4589bf618b6f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.130904] env[65385]: DEBUG oslo_vmware.api [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454291, 'name': PowerOnVM_Task, 'duration_secs': 0.588001} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.131331] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1047.131555] env[65385]: INFO nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Took 7.44 seconds to spawn the instance on the hypervisor. [ 1047.131739] env[65385]: DEBUG nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1047.132992] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7e545c-0e6d-4e96-ba68-6cd4a151c5c5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.346577] env[65385]: WARNING neutronclient.v2_0.client [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.371479] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.383693] env[65385]: WARNING neutronclient.v2_0.client [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.383798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.383934] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquired lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.384079] env[65385]: DEBUG nova.network.neutron [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1047.384754] env[65385]: DEBUG nova.objects.instance [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'info_cache' on Instance uuid 8a974e99-a7b1-438e-aace-9ae82352495c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.402337] env[65385]: INFO nova.scheduler.client.report [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted allocations for instance 537393de-ea48-4d34-8e04-95331bbc099d [ 1047.574279] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b87b2de3-865d-416a-885e-0957783de940 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.024s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.651310] env[65385]: INFO nova.compute.manager [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Took 12.41 seconds to build instance. [ 1047.909489] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ee353049-5119-4492-834c-9d56e94aeb75 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "537393de-ea48-4d34-8e04-95331bbc099d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.407s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.001461] env[65385]: DEBUG nova.network.neutron [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Successfully updated port: 0fc23e28-1265-46b7-aa43-fd72e2269330 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1048.153324] env[65385]: DEBUG oslo_concurrency.lockutils [None req-689e2cea-e03b-49af-b2e7-c38df1b54d72 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e13d7fec-250a-412e-b952-f189214c8c1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.923s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.391400] env[65385]: WARNING neutronclient.v2_0.client [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1048.392335] env[65385]: WARNING openstack [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.392815] env[65385]: WARNING openstack [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.501919] env[65385]: WARNING openstack [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.502383] env[65385]: WARNING openstack [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.509338] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-77d718c8-4959-483a-8717-c00f2be1d0bb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.509511] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-77d718c8-4959-483a-8717-c00f2be1d0bb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.509699] env[65385]: DEBUG nova.network.neutron [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1048.573715] env[65385]: WARNING neutronclient.v2_0.client [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1048.574543] env[65385]: WARNING openstack [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.575773] env[65385]: WARNING openstack [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.677478] env[65385]: DEBUG nova.network.neutron [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance_info_cache with network_info: [{"id": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "address": "fa:16:3e:1a:55:36", "network": {"id": "22965a88-dbed-4a4f-9ad4-7996c38be45f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1618171652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "754b1facaaa14501b2204c98e1d7a5a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26f849c-40", "ovs_interfaceid": "f26f849c-4075-4ee1-b174-88ce8e66b69c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1049.013734] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.014505] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.056854] env[65385]: DEBUG nova.network.neutron [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1049.074988] env[65385]: DEBUG nova.compute.manager [req-2f3f19ac-8b1f-4ece-bbf4-e48730942601 req-fbb2bc23-cf8d-4bc9-8f9e-2439d4850a70 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Received event network-vif-plugged-0fc23e28-1265-46b7-aa43-fd72e2269330 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1049.075402] env[65385]: DEBUG oslo_concurrency.lockutils [req-2f3f19ac-8b1f-4ece-bbf4-e48730942601 req-fbb2bc23-cf8d-4bc9-8f9e-2439d4850a70 service nova] Acquiring lock "77d718c8-4959-483a-8717-c00f2be1d0bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.075618] env[65385]: DEBUG oslo_concurrency.lockutils [req-2f3f19ac-8b1f-4ece-bbf4-e48730942601 req-fbb2bc23-cf8d-4bc9-8f9e-2439d4850a70 service nova] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.075794] env[65385]: DEBUG oslo_concurrency.lockutils [req-2f3f19ac-8b1f-4ece-bbf4-e48730942601 req-fbb2bc23-cf8d-4bc9-8f9e-2439d4850a70 service nova] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.075971] env[65385]: DEBUG nova.compute.manager [req-2f3f19ac-8b1f-4ece-bbf4-e48730942601 req-fbb2bc23-cf8d-4bc9-8f9e-2439d4850a70 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] No waiting events found dispatching network-vif-plugged-0fc23e28-1265-46b7-aa43-fd72e2269330 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1049.076150] env[65385]: WARNING nova.compute.manager [req-2f3f19ac-8b1f-4ece-bbf4-e48730942601 req-fbb2bc23-cf8d-4bc9-8f9e-2439d4850a70 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Received unexpected event network-vif-plugged-0fc23e28-1265-46b7-aa43-fd72e2269330 for instance with vm_state building and task_state spawning. [ 1049.082565] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.083063] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.153591] env[65385]: WARNING neutronclient.v2_0.client [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.154370] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.154765] env[65385]: WARNING openstack [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.179947] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Releasing lock "refresh_cache-8a974e99-a7b1-438e-aace-9ae82352495c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.180261] env[65385]: DEBUG nova.objects.instance [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lazy-loading 'migration_context' on Instance uuid 8a974e99-a7b1-438e-aace-9ae82352495c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.248306] env[65385]: DEBUG nova.network.neutron [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Updating instance_info_cache with network_info: [{"id": "0fc23e28-1265-46b7-aa43-fd72e2269330", "address": "fa:16:3e:d0:1a:e5", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fc23e28-12", "ovs_interfaceid": "0fc23e28-1265-46b7-aa43-fd72e2269330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1049.683276] env[65385]: DEBUG nova.objects.base [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Object Instance<8a974e99-a7b1-438e-aace-9ae82352495c> lazy-loaded attributes: info_cache,migration_context {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1049.684627] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e22f0ae-90c1-4463-ba1f-a385acfd13f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.704751] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c680409-72ef-443b-9676-d04d0085e45e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.711890] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1049.711890] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525fc716-9b85-88f9-1b9e-acdd5fb51a43" [ 1049.711890] env[65385]: _type = "Task" [ 1049.711890] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.721951] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525fc716-9b85-88f9-1b9e-acdd5fb51a43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.751803] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-77d718c8-4959-483a-8717-c00f2be1d0bb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.752233] env[65385]: DEBUG nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Instance network_info: |[{"id": "0fc23e28-1265-46b7-aa43-fd72e2269330", "address": "fa:16:3e:d0:1a:e5", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fc23e28-12", "ovs_interfaceid": "0fc23e28-1265-46b7-aa43-fd72e2269330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1049.752728] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:1a:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0fc23e28-1265-46b7-aa43-fd72e2269330', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.761680] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1049.762070] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1049.762320] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5fec21f8-954c-4366-a9fe-148c3b15b2a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.784755] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.784755] env[65385]: value = "task-4454293" [ 1049.784755] env[65385]: _type = "Task" [ 1049.784755] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.795692] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454293, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.858632] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.858957] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.859207] env[65385]: DEBUG nova.compute.manager [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Going to confirm migration 5 {{(pid=65385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1050.010090] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "e13d7fec-250a-412e-b952-f189214c8c1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.010358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e13d7fec-250a-412e-b952-f189214c8c1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.010646] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "e13d7fec-250a-412e-b952-f189214c8c1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.011023] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e13d7fec-250a-412e-b952-f189214c8c1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.011217] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e13d7fec-250a-412e-b952-f189214c8c1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.013348] env[65385]: INFO nova.compute.manager [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Terminating instance [ 1050.222696] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525fc716-9b85-88f9-1b9e-acdd5fb51a43, 'name': SearchDatastore_Task, 'duration_secs': 0.013798} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.222987] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.223255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.295616] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454293, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.365444] env[65385]: WARNING neutronclient.v2_0.client [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.401741] env[65385]: WARNING neutronclient.v2_0.client [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.402144] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.402310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.402482] env[65385]: DEBUG nova.network.neutron [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1050.402658] env[65385]: DEBUG nova.objects.instance [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'info_cache' on Instance uuid 59f26752-75ab-45e5-b0a9-bbcef44ecd7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.459321] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.459606] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.517073] env[65385]: DEBUG nova.compute.manager [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1050.517316] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1050.520418] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cc3eff-6160-4c5a-a2a7-1106688ea6f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.528727] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1050.529095] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b35dd2b-3a86-45df-8be2-948df77e480b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.537867] env[65385]: DEBUG oslo_vmware.api [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1050.537867] env[65385]: value = "task-4454294" [ 1050.537867] env[65385]: _type = "Task" [ 1050.537867] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.547109] env[65385]: DEBUG oslo_vmware.api [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.800604] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454293, 'name': CreateVM_Task, 'duration_secs': 0.785263} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.800801] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.801366] env[65385]: WARNING neutronclient.v2_0.client [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.801738] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.801968] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.802538] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1050.805730] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88a6e073-8816-43bf-964e-b9efd626185c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.813018] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1050.813018] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5285c893-c405-d048-9302-ccacb61be041" [ 1050.813018] env[65385]: _type = "Task" [ 1050.813018] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.821931] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5285c893-c405-d048-9302-ccacb61be041, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.909967] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed69b291-1957-4003-973f-3fc2ccf95317 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.918424] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f027ca4-d8b1-457a-b694-02ec93559df6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.953886] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6607a27b-d4c9-486c-9865-d2ba87419f37 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.962286] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c663dba-d710-47b2-9aa3-07a31f5d54a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.966622] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.966797] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.967817] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9de16e-a7d7-4e14-9577-bbbf90f9eda4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.981594] env[65385]: DEBUG nova.compute.provider_tree [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.999412] env[65385]: DEBUG nova.scheduler.client.report [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1051.003043] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae56b9c2-60d1-41a0-bf66-a4ea24de9707 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.029792] env[65385]: WARNING neutronclient.v2_0.client [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.035464] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfiguring VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1051.036088] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-322ed3cb-5d6d-4c83-a09d-85b49fd30986 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.060201] env[65385]: DEBUG oslo_vmware.api [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454294, 'name': PowerOffVM_Task, 'duration_secs': 0.399582} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.061686] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1051.061888] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1051.062236] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1051.062236] env[65385]: value = "task-4454295" [ 1051.062236] env[65385]: _type = "Task" [ 1051.062236] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.062444] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1210c5ce-c227-4e29-a0b4-e8f7c7ba7f95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.072909] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.136945] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1051.137579] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1051.137579] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleting the datastore file [datastore2] e13d7fec-250a-412e-b952-f189214c8c1e {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1051.137773] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe7f1557-7a04-47eb-b309-04ceb40d69dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.146461] env[65385]: DEBUG oslo_vmware.api [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1051.146461] env[65385]: value = "task-4454297" [ 1051.146461] env[65385]: _type = "Task" [ 1051.146461] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.155326] env[65385]: DEBUG oslo_vmware.api [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.170903] env[65385]: DEBUG nova.compute.manager [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Received event network-changed-0fc23e28-1265-46b7-aa43-fd72e2269330 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1051.171176] env[65385]: DEBUG nova.compute.manager [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Refreshing instance network info cache due to event network-changed-0fc23e28-1265-46b7-aa43-fd72e2269330. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1051.171364] env[65385]: DEBUG oslo_concurrency.lockutils [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Acquiring lock "refresh_cache-77d718c8-4959-483a-8717-c00f2be1d0bb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.171504] env[65385]: DEBUG oslo_concurrency.lockutils [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Acquired lock "refresh_cache-77d718c8-4959-483a-8717-c00f2be1d0bb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1051.171659] env[65385]: DEBUG nova.network.neutron [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Refreshing network info cache for port 0fc23e28-1265-46b7-aa43-fd72e2269330 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1051.249966] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquiring lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.250333] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1051.327449] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5285c893-c405-d048-9302-ccacb61be041, 'name': SearchDatastore_Task, 'duration_secs': 0.021621} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.327834] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.327972] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.328227] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.328368] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1051.328569] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.329223] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-deffe47b-1190-48c6-b97f-bb8342faa81b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.337742] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.337896] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.338630] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f8cdcea-ca0f-4388-a7e5-b6178b3249fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.344594] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1051.344594] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c9b222-7e09-b2fb-0aed-6e874da7a151" [ 1051.344594] env[65385]: _type = "Task" [ 1051.344594] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.352502] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c9b222-7e09-b2fb-0aed-6e874da7a151, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.408864] env[65385]: WARNING neutronclient.v2_0.client [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.409506] env[65385]: WARNING openstack [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.409847] env[65385]: WARNING openstack [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.575395] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.657942] env[65385]: DEBUG oslo_vmware.api [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183879} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.658217] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.658468] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1051.658668] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1051.658845] env[65385]: INFO nova.compute.manager [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1051.659139] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1051.659339] env[65385]: DEBUG nova.compute.manager [-] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1051.659475] env[65385]: DEBUG nova.network.neutron [-] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1051.659696] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.660367] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.660626] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.676468] env[65385]: WARNING neutronclient.v2_0.client [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.677111] env[65385]: WARNING openstack [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.677523] env[65385]: WARNING openstack [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.700683] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.752779] env[65385]: DEBUG nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1051.855867] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c9b222-7e09-b2fb-0aed-6e874da7a151, 'name': SearchDatastore_Task, 'duration_secs': 0.010992} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.856716] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50fc46e2-b2f5-4617-863c-2ccdcadf3836 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.862970] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1051.862970] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5255815d-4053-e1de-8c26-5308619758f8" [ 1051.862970] env[65385]: _type = "Task" [ 1051.862970] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.871304] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5255815d-4053-e1de-8c26-5308619758f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.973270] env[65385]: WARNING openstack [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.973665] env[65385]: WARNING openstack [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.017080] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 1.793s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.038878] env[65385]: WARNING openstack [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.039224] env[65385]: WARNING openstack [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.065727] env[65385]: WARNING neutronclient.v2_0.client [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1052.066373] env[65385]: WARNING openstack [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.066720] env[65385]: WARNING openstack [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.084641] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.134033] env[65385]: WARNING neutronclient.v2_0.client [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1052.134820] env[65385]: WARNING openstack [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.135368] env[65385]: WARNING openstack [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.147983] env[65385]: DEBUG nova.compute.manager [req-4b09f938-1129-4f13-9584-be65e8e07651 req-e523e302-d50a-4fdb-8056-8d6e7ab7fc5c service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Received event network-vif-deleted-3f844c95-3619-44bb-945c-bec15b1f30ec {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1052.148185] env[65385]: INFO nova.compute.manager [req-4b09f938-1129-4f13-9584-be65e8e07651 req-e523e302-d50a-4fdb-8056-8d6e7ab7fc5c service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Neutron deleted interface 3f844c95-3619-44bb-945c-bec15b1f30ec; detaching it from the instance and deleting it from the info cache [ 1052.148346] env[65385]: DEBUG nova.network.neutron [req-4b09f938-1129-4f13-9584-be65e8e07651 req-e523e302-d50a-4fdb-8056-8d6e7ab7fc5c service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.154132] env[65385]: DEBUG nova.network.neutron [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [{"id": "59289962-8634-414f-9bd2-b3f5b19af9b4", "address": "fa:16:3e:8a:a4:1c", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59289962-86", "ovs_interfaceid": "59289962-8634-414f-9bd2-b3f5b19af9b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.225223] env[65385]: DEBUG nova.network.neutron [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Updated VIF entry in instance network info cache for port 0fc23e28-1265-46b7-aa43-fd72e2269330. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1052.225223] env[65385]: DEBUG nova.network.neutron [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Updating instance_info_cache with network_info: [{"id": "0fc23e28-1265-46b7-aa43-fd72e2269330", "address": "fa:16:3e:d0:1a:e5", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fc23e28-12", "ovs_interfaceid": "0fc23e28-1265-46b7-aa43-fd72e2269330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.278977] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.279282] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.281364] env[65385]: INFO nova.compute.claims [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.375427] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5255815d-4053-e1de-8c26-5308619758f8, 'name': SearchDatastore_Task, 'duration_secs': 0.013397} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.375676] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1052.375922] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1052.376195] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63c52024-4fd1-4351-bb18-bc3221175737 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.383679] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1052.383679] env[65385]: value = "task-4454298" [ 1052.383679] env[65385]: _type = "Task" [ 1052.383679] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.392232] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454298, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.482937] env[65385]: DEBUG nova.network.neutron [-] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.584784] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.598490] env[65385]: INFO nova.scheduler.client.report [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted allocation for migration 64929b85-d30a-41d5-9d93-bceef0b8938f [ 1052.655518] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-59f26752-75ab-45e5-b0a9-bbcef44ecd7f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1052.655785] env[65385]: DEBUG nova.objects.instance [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'migration_context' on Instance uuid 59f26752-75ab-45e5-b0a9-bbcef44ecd7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.656816] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9393eae4-e253-49c1-8225-7df75ec2f03b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.667431] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a1f84f-a54e-4bea-8f31-339171152b3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.699252] env[65385]: DEBUG nova.compute.manager [req-4b09f938-1129-4f13-9584-be65e8e07651 req-e523e302-d50a-4fdb-8056-8d6e7ab7fc5c service nova] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Detach interface failed, port_id=3f844c95-3619-44bb-945c-bec15b1f30ec, reason: Instance e13d7fec-250a-412e-b952-f189214c8c1e could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1052.728169] env[65385]: DEBUG oslo_concurrency.lockutils [req-1955d097-031c-4680-8f02-18ecb22a32c1 req-9c9b296f-51a0-42b9-8f50-21d4979e0ec5 service nova] Releasing lock "refresh_cache-77d718c8-4959-483a-8717-c00f2be1d0bb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1052.895998] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454298, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.986116] env[65385]: INFO nova.compute.manager [-] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Took 1.33 seconds to deallocate network for instance. [ 1053.086095] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.105108] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.266s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1053.160390] env[65385]: DEBUG nova.objects.base [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Object Instance<59f26752-75ab-45e5-b0a9-bbcef44ecd7f> lazy-loaded attributes: info_cache,migration_context {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1053.161404] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0b2e24-ed6f-42dd-9a5c-3c5bdc9cc910 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.184125] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35ab5ab9-1174-416d-a80d-03b161845101 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.190568] env[65385]: DEBUG oslo_vmware.api [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1053.190568] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5278e9fb-2c8c-5595-044a-3518d97b86cb" [ 1053.190568] env[65385]: _type = "Task" [ 1053.190568] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.200409] env[65385]: DEBUG oslo_vmware.api [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5278e9fb-2c8c-5595-044a-3518d97b86cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.396994] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454298, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.758978} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.397334] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1053.397463] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.397707] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0baa217f-7ba1-4e01-bd05-1f9283e30109 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.404859] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1053.404859] env[65385]: value = "task-4454299" [ 1053.404859] env[65385]: _type = "Task" [ 1053.404859] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.416750] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454299, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.455775] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0218fca-4bd7-4b72-8af9-36c7ca03a41d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.463711] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f55190-eeeb-4846-a534-467cb2bd3ae3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.494924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1053.496400] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637ba08d-b0a4-48fc-84cc-39140b78eb15 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.505419] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7d21e8-59a0-4c5b-9423-3f8b98183fa0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.520033] env[65385]: DEBUG nova.compute.provider_tree [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.587154] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.700949] env[65385]: DEBUG oslo_vmware.api [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5278e9fb-2c8c-5595-044a-3518d97b86cb, 'name': SearchDatastore_Task, 'duration_secs': 0.009981} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.701218] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.638878] env[65385]: DEBUG nova.scheduler.client.report [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1054.654945] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.657893] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454299, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06671} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.658694] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1054.659463] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71baf5ae-597e-4754-9bf1-c21a4b695436 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.686764] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1054.687692] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60adbcf2-d3d1-4a5b-9155-081228684212 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.709258] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1054.709258] env[65385]: value = "task-4454300" [ 1054.709258] env[65385]: _type = "Task" [ 1054.709258] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.718152] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454300, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.138874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "8a974e99-a7b1-438e-aace-9ae82352495c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.139184] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.139405] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.139582] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.139743] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.142332] env[65385]: INFO nova.compute.manager [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Terminating instance [ 1055.144276] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.865s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.144673] env[65385]: DEBUG nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1055.151148] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.656s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.152049] env[65385]: DEBUG nova.objects.instance [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'resources' on Instance uuid e13d7fec-250a-412e-b952-f189214c8c1e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.159213] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.219019] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454300, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.656486] env[65385]: DEBUG nova.compute.utils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1055.658415] env[65385]: DEBUG nova.compute.manager [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1055.658735] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.661354] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.662189] env[65385]: DEBUG nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1055.662524] env[65385]: DEBUG nova.network.neutron [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1055.662868] env[65385]: WARNING neutronclient.v2_0.client [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.662921] env[65385]: WARNING neutronclient.v2_0.client [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.663455] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.663790] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.670583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453b9861-56a0-48c3-a400-68c5d82f2620 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.680994] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.683859] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d907125-d3b0-4ebb-a654-2703fb89c31e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.692798] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1055.692798] env[65385]: value = "task-4454301" [ 1055.692798] env[65385]: _type = "Task" [ 1055.692798] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.704397] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.723552] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454300, 'name': ReconfigVM_Task, 'duration_secs': 0.535484} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.723552] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1055.724945] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2bd5144b-d597-4633-a583-bab6fd8cbea4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.731911] env[65385]: DEBUG nova.policy [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a2a0727bc2643fe908c2fafcdd88994', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87efee8d30ee4a69bdd8dc784e809bfe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1055.738641] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1055.738641] env[65385]: value = "task-4454302" [ 1055.738641] env[65385]: _type = "Task" [ 1055.738641] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.751652] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454302, 'name': Rename_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.852791] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fff5839-4198-411e-9c48-5fef81b8e522 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.861924] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1589c359-a912-43b6-9376-fecdf7d29cd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.894612] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f633f944-d587-441d-aa30-c3c165fe9d97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.902843] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6075fa56-304c-43ad-b9e8-9bbd7581d79e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.918389] env[65385]: DEBUG nova.compute.provider_tree [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.040608] env[65385]: DEBUG nova.network.neutron [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Successfully created port: b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1056.155119] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.163286] env[65385]: DEBUG nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1056.210028] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454301, 'name': PowerOffVM_Task, 'duration_secs': 0.21954} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.210028] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.210028] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.210028] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af865fda-1537-4cf5-8a76-afae8a67448f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.250011] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454302, 'name': Rename_Task, 'duration_secs': 0.161749} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.250504] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1056.250673] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfcff6bb-4e23-4c29-aa5f-d8a120716cd8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.260316] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1056.260316] env[65385]: value = "task-4454304" [ 1056.260316] env[65385]: _type = "Task" [ 1056.260316] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.269412] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454304, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.280579] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.280791] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.280992] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleting the datastore file [datastore1] 8a974e99-a7b1-438e-aace-9ae82352495c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.281272] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7768ef5e-a84b-4ff0-b4cb-c3c39c76e3e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.288438] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for the task: (returnval){ [ 1056.288438] env[65385]: value = "task-4454305" [ 1056.288438] env[65385]: _type = "Task" [ 1056.288438] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.301124] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454305, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.422184] env[65385]: DEBUG nova.scheduler.client.report [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1056.659922] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.669653] env[65385]: INFO nova.virt.block_device [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Booting with volume aee71a50-6144-4280-9b67-ff58bba61633 at /dev/sda [ 1056.719565] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a557640b-51fc-477d-8d4a-faabeae01649 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.732225] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ced8d6-d641-4969-907a-481fa1119680 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.770289] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2849fa47-a91a-417e-b5b4-d463b9e7a646 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.778368] env[65385]: DEBUG oslo_vmware.api [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454304, 'name': PowerOnVM_Task, 'duration_secs': 0.496996} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.779889] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1056.779889] env[65385]: INFO nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Took 9.74 seconds to spawn the instance on the hypervisor. [ 1056.780065] env[65385]: DEBUG nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1056.780863] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4462f34-97ed-4552-98ce-1e355eb153be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.786008] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de728262-3130-47e2-8095-639e5d888ef9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.809920] env[65385]: DEBUG oslo_vmware.api [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Task: {'id': task-4454305, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184881} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.810212] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1056.810383] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1056.810549] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1056.810712] env[65385]: INFO nova.compute.manager [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1056.810950] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1056.822534] env[65385]: DEBUG nova.compute.manager [-] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1056.822657] env[65385]: DEBUG nova.network.neutron [-] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1056.822957] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.823515] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.823795] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.831578] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5247eeb7-2e8b-4946-8d3f-3ed3d05261d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.839318] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b74465-9872-4e5b-a907-bc82656c2e45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.844672] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.855280] env[65385]: DEBUG nova.virt.block_device [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Updating existing volume attachment record: 9c0b0fb7-07a5-419a-b52c-1ae4f825faab {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1056.866580] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.927544] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.930938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.230s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.950655] env[65385]: INFO nova.scheduler.client.report [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted allocations for instance e13d7fec-250a-412e-b952-f189214c8c1e [ 1057.123727] env[65385]: DEBUG nova.compute.manager [req-ff33ca23-7007-446d-9df2-081447a68cb5 req-32d8e569-79ec-4f0b-a4a8-444d39a8c1e8 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Received event network-vif-deleted-f26f849c-4075-4ee1-b174-88ce8e66b69c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1057.123986] env[65385]: INFO nova.compute.manager [req-ff33ca23-7007-446d-9df2-081447a68cb5 req-32d8e569-79ec-4f0b-a4a8-444d39a8c1e8 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Neutron deleted interface f26f849c-4075-4ee1-b174-88ce8e66b69c; detaching it from the instance and deleting it from the info cache [ 1057.124241] env[65385]: DEBUG nova.network.neutron [req-ff33ca23-7007-446d-9df2-081447a68cb5 req-32d8e569-79ec-4f0b-a4a8-444d39a8c1e8 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.157943] env[65385]: DEBUG oslo_vmware.api [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454295, 'name': ReconfigVM_Task, 'duration_secs': 5.858756} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.158444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.158655] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Reconfigured VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1057.159114] env[65385]: WARNING neutronclient.v2_0.client [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.159434] env[65385]: WARNING neutronclient.v2_0.client [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.159985] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.160377] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.191786] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.192054] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.314700] env[65385]: INFO nova.compute.manager [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Took 14.22 seconds to build instance. [ 1057.453303] env[65385]: DEBUG nova.compute.manager [req-c12a6200-4f2c-4bc0-9f3a-6b78b121ade0 req-a71fe518-c89d-4d94-abc2-deeb3669f3ee service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Received event network-vif-plugged-b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1057.453490] env[65385]: DEBUG oslo_concurrency.lockutils [req-c12a6200-4f2c-4bc0-9f3a-6b78b121ade0 req-a71fe518-c89d-4d94-abc2-deeb3669f3ee service nova] Acquiring lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.453690] env[65385]: DEBUG oslo_concurrency.lockutils [req-c12a6200-4f2c-4bc0-9f3a-6b78b121ade0 req-a71fe518-c89d-4d94-abc2-deeb3669f3ee service nova] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.453867] env[65385]: DEBUG oslo_concurrency.lockutils [req-c12a6200-4f2c-4bc0-9f3a-6b78b121ade0 req-a71fe518-c89d-4d94-abc2-deeb3669f3ee service nova] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.454233] env[65385]: DEBUG nova.compute.manager [req-c12a6200-4f2c-4bc0-9f3a-6b78b121ade0 req-a71fe518-c89d-4d94-abc2-deeb3669f3ee service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] No waiting events found dispatching network-vif-plugged-b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1057.454422] env[65385]: WARNING nova.compute.manager [req-c12a6200-4f2c-4bc0-9f3a-6b78b121ade0 req-a71fe518-c89d-4d94-abc2-deeb3669f3ee service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Received unexpected event network-vif-plugged-b9948bb7-a208-446d-a45e-2a74affe299b for instance with vm_state building and task_state block_device_mapping. [ 1057.460246] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8540c384-a475-48f3-b54a-63db04b3d517 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e13d7fec-250a-412e-b952-f189214c8c1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.450s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.562888] env[65385]: DEBUG nova.network.neutron [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Successfully updated port: b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1057.603399] env[65385]: DEBUG nova.network.neutron [-] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.618747] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0278c2c-be3d-4382-a231-6a433886604d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.627503] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ad588a3-cbeb-49eb-8fb8-84ddb41f802f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.630013] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a991c78-52f8-4467-8716-4eca498d4830 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.663748] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02f4d51-083e-413c-868d-4685a4e5bc34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.668760] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6cbc34-7ab2-40cb-81b6-682fbd2dcb7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.685255] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8a23c3-0975-4103-9395-f2a742b4f69a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.701128] env[65385]: DEBUG nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1057.707026] env[65385]: DEBUG nova.compute.manager [req-ff33ca23-7007-446d-9df2-081447a68cb5 req-32d8e569-79ec-4f0b-a4a8-444d39a8c1e8 service nova] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Detach interface failed, port_id=f26f849c-4075-4ee1-b174-88ce8e66b69c, reason: Instance 8a974e99-a7b1-438e-aace-9ae82352495c could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1057.712310] env[65385]: DEBUG nova.compute.provider_tree [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.816743] env[65385]: DEBUG oslo_concurrency.lockutils [None req-06b0b5db-77d2-4c19-a40c-9d98a02881e1 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.730s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.843733] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.843944] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1058.066964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquiring lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.068501] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquired lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.068501] env[65385]: DEBUG nova.network.neutron [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1058.109669] env[65385]: INFO nova.compute.manager [-] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Took 1.29 seconds to deallocate network for instance. [ 1058.216483] env[65385]: DEBUG nova.scheduler.client.report [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1058.238339] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.421719] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.422044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.422322] env[65385]: DEBUG nova.network.neutron [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1058.572474] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.572912] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.611577] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.611814] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.612103] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.612256] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.612445] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.614706] env[65385]: INFO nova.compute.manager [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Terminating instance [ 1058.619363] env[65385]: DEBUG nova.network.neutron [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1058.622782] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.641599] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.641976] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.703616] env[65385]: WARNING neutronclient.v2_0.client [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.704406] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.704819] env[65385]: WARNING openstack [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.805607] env[65385]: DEBUG nova.network.neutron [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Updating instance_info_cache with network_info: [{"id": "b9948bb7-a208-446d-a45e-2a74affe299b", "address": "fa:16:3e:87:22:22", "network": {"id": "006c9946-474b-48b9-8af4-c5ae5034f3d6", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2098203597-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87efee8d30ee4a69bdd8dc784e809bfe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9948bb7-a2", "ovs_interfaceid": "b9948bb7-a208-446d-a45e-2a74affe299b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1058.843361] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1058.843788] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1058.871567] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "04ebe83c-cab5-45e1-9f5f-018481f63912" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.871794] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.925190] env[65385]: WARNING neutronclient.v2_0.client [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.925895] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.926341] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.956627] env[65385]: DEBUG nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1058.957144] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1058.957361] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1058.957507] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1058.957679] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1058.957816] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1058.957953] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1058.958165] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.958315] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1058.958474] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1058.958621] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1058.958786] env[65385]: DEBUG nova.virt.hardware [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1058.961878] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd89141-a00e-42a1-a960-dd4bcfd79567 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.973592] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8ce860-5d3d-4a3a-8d08-e96e9193b5b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.042130] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.042575] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.104734] env[65385]: WARNING neutronclient.v2_0.client [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.105402] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.105747] env[65385]: WARNING openstack [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.117894] env[65385]: DEBUG nova.compute.manager [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1059.118166] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.118979] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c1e7cba-d2f8-4159-bf1b-978f2b58f53c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.129857] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.129857] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bdee80d5-20d4-4281-9353-ed8fd97bbbcc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.136855] env[65385]: DEBUG oslo_vmware.api [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1059.136855] env[65385]: value = "task-4454306" [ 1059.136855] env[65385]: _type = "Task" [ 1059.136855] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.147933] env[65385]: DEBUG oslo_vmware.api [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454306, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.167413] env[65385]: DEBUG nova.compute.manager [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-vif-deleted-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1059.167413] env[65385]: INFO nova.compute.manager [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Neutron deleted interface ef6b6d8f-dd4a-4682-991b-9c6dd0d27525; detaching it from the instance and deleting it from the info cache [ 1059.167730] env[65385]: DEBUG nova.network.neutron [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "05d2fd35-d0a1-44d7-9944-0b6445e2bd9b", "address": "fa:16:3e:2f:de:4d", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05d2fd35-d0", "ovs_interfaceid": "05d2fd35-d0a1-44d7-9944-0b6445e2bd9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.227184] env[65385]: INFO nova.network.neutron [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Port ef6b6d8f-dd4a-4682-991b-9c6dd0d27525 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1059.227642] env[65385]: INFO nova.network.neutron [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Port 05d2fd35-d0a1-44d7-9944-0b6445e2bd9b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1059.228170] env[65385]: DEBUG nova.network.neutron [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.231447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.301s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1059.237641] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.998s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.239111] env[65385]: INFO nova.compute.claims [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.309565] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Releasing lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.310019] env[65385]: DEBUG nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance network_info: |[{"id": "b9948bb7-a208-446d-a45e-2a74affe299b", "address": "fa:16:3e:87:22:22", "network": {"id": "006c9946-474b-48b9-8af4-c5ae5034f3d6", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2098203597-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87efee8d30ee4a69bdd8dc784e809bfe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9948bb7-a2", "ovs_interfaceid": "b9948bb7-a208-446d-a45e-2a74affe299b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1059.311105] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:22:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9948bb7-a208-446d-a45e-2a74affe299b', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.319669] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Creating folder: Project (87efee8d30ee4a69bdd8dc784e809bfe). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1059.320108] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d9bdfd11-3268-4fde-8d70-e9dfc4221f09 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.335794] env[65385]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1059.336141] env[65385]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65385) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1059.336792] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Folder already exists: Project (87efee8d30ee4a69bdd8dc784e809bfe). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1059.336792] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Creating folder: Instances. Parent ref: group-v871153. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1059.336957] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a16aed94-6f28-4093-ad80-52e8d502b6c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.347015] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.349301] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Created folder: Instances in parent group-v871153. [ 1059.349654] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1059.349944] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.350127] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fab3e36d-f5e6-48c2-b080-beab46668fd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.374458] env[65385]: DEBUG nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1059.377619] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.377619] env[65385]: value = "task-4454309" [ 1059.377619] env[65385]: _type = "Task" [ 1059.377619] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.384053] env[65385]: INFO nova.compute.manager [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Rebuilding instance [ 1059.391167] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454309, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.438602] env[65385]: DEBUG nova.compute.manager [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1059.439560] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12514dee-822f-4bfd-8327-eec3f2685bf3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.483268] env[65385]: DEBUG nova.compute.manager [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Received event network-changed-b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1059.483471] env[65385]: DEBUG nova.compute.manager [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Refreshing instance network info cache due to event network-changed-b9948bb7-a208-446d-a45e-2a74affe299b. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1059.483704] env[65385]: DEBUG oslo_concurrency.lockutils [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Acquiring lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.483843] env[65385]: DEBUG oslo_concurrency.lockutils [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Acquired lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1059.484453] env[65385]: DEBUG nova.network.neutron [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Refreshing network info cache for port b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1059.648161] env[65385]: DEBUG oslo_vmware.api [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454306, 'name': PowerOffVM_Task, 'duration_secs': 0.222324} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.648537] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.648707] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.648969] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7463995-bab2-48c3-8e73-1f3e15c603e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.671825] env[65385]: DEBUG oslo_concurrency.lockutils [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] Acquiring lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.738039] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.772150] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.772498] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.772822] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleting the datastore file [datastore2] a5fe2ad9-6dfe-485b-bf27-beb218d31467 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.773246] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4cdfe8a-37d8-4531-96d3-2f08ef9b4957 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.782363] env[65385]: DEBUG oslo_vmware.api [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1059.782363] env[65385]: value = "task-4454311" [ 1059.782363] env[65385]: _type = "Task" [ 1059.782363] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.791770] env[65385]: DEBUG oslo_vmware.api [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454311, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.817493] env[65385]: INFO nova.scheduler.client.report [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted allocation for migration e3a9f52b-e795-4f9a-92e7-8b80ef42a93d [ 1059.895614] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454309, 'name': CreateVM_Task, 'duration_secs': 0.399757} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.895822] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1059.896445] env[65385]: WARNING neutronclient.v2_0.client [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.896805] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871157', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'name': 'volume-aee71a50-6144-4280-9b67-ff58bba61633', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b1abf6ab-a3d5-4078-9568-68db0a7f03cc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'serial': 'aee71a50-6144-4280-9b67-ff58bba61633'}, 'device_type': None, 'guest_format': None, 'attachment_id': '9c0b0fb7-07a5-419a-b52c-1ae4f825faab', 'mount_device': '/dev/sda', 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=65385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1059.897043] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Root volume attach. Driver type: vmdk {{(pid=65385) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1059.897892] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d0b500-cb02-4557-acc1-f61523b0e492 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.901427] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.908192] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd048a2-ca23-4dc9-a9a3-464e162eb706 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.915168] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d106b612-8a62-4e80-8b9b-280e74e1ec22 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.921390] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-5756eecc-7fb1-4f2d-92d3-2972fd5ee959 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.931724] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1059.931724] env[65385]: value = "task-4454312" [ 1059.931724] env[65385]: _type = "Task" [ 1059.931724] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.941226] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.987600] env[65385]: WARNING neutronclient.v2_0.client [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.988366] env[65385]: WARNING openstack [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.988653] env[65385]: WARNING openstack [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.119446] env[65385]: WARNING openstack [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.119852] env[65385]: WARNING openstack [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.190107] env[65385]: WARNING neutronclient.v2_0.client [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.190798] env[65385]: WARNING openstack [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.191161] env[65385]: WARNING openstack [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.244388] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6e42ca60-0279-4a08-871c-de4082aa0131 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-a5fe2ad9-6dfe-485b-bf27-beb218d31467-ef6b6d8f-dd4a-4682-991b-9c6dd0d27525" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.785s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.294432] env[65385]: DEBUG oslo_vmware.api [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454311, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.303881} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.294725] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.294912] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.295103] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.295347] env[65385]: INFO nova.compute.manager [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1060.295609] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1060.295814] env[65385]: DEBUG nova.compute.manager [-] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1060.295913] env[65385]: DEBUG nova.network.neutron [-] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1060.296230] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.297034] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.297216] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.308929] env[65385]: DEBUG nova.network.neutron [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Updated VIF entry in instance network info cache for port b9948bb7-a208-446d-a45e-2a74affe299b. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1060.309326] env[65385]: DEBUG nova.network.neutron [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Updating instance_info_cache with network_info: [{"id": "b9948bb7-a208-446d-a45e-2a74affe299b", "address": "fa:16:3e:87:22:22", "network": {"id": "006c9946-474b-48b9-8af4-c5ae5034f3d6", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2098203597-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87efee8d30ee4a69bdd8dc784e809bfe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9948bb7-a2", "ovs_interfaceid": "b9948bb7-a208-446d-a45e-2a74affe299b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1060.323347] env[65385]: DEBUG oslo_concurrency.lockutils [None req-85bcdfe7-c529-4ca3-97f5-8ed9b0546493 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.464s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.382322] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.407978] env[65385]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 05d2fd35-d0a1-44d7-9944-0b6445e2bd9b could not be found.", "detail": ""}} {{(pid=65385) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:265}} [ 1060.408233] env[65385]: DEBUG nova.network.neutron [-] Unable to show port 05d2fd35-d0a1-44d7-9944-0b6445e2bd9b as it no longer exists. {{(pid=65385) _unbind_ports /opt/stack/nova/nova/network/neutron.py:700}} [ 1060.446669] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 34%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.455884] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1060.456803] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1769af93-2463-4deb-96d8-d21679b6307c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.465931] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1060.465931] env[65385]: value = "task-4454313" [ 1060.465931] env[65385]: _type = "Task" [ 1060.465931] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.481188] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.487436] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62b0349-38ec-48ca-91fe-856b0e73269b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.497221] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c266f8b6-3112-492d-93c9-49b35f24b950 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.535285] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119d3a8f-f578-4e97-a3d7-0f3055369e2c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.546316] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8501351-e4b5-4ed2-842f-9bdc228b2e54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.568762] env[65385]: DEBUG nova.compute.provider_tree [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.812669] env[65385]: DEBUG oslo_concurrency.lockutils [req-9149e4b5-cdad-4214-9195-6079d0612368 req-2649d5ba-b32d-4625-8bb5-412e5d7dbc08 service nova] Releasing lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1060.821674] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.822020] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.870022] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.870378] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.870597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.870777] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.870945] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.873576] env[65385]: INFO nova.compute.manager [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Terminating instance [ 1060.946432] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 47%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.976470] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454313, 'name': PowerOffVM_Task, 'duration_secs': 0.286809} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.976796] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1060.976969] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1060.977916] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17cbad6-f0bd-4d52-9f30-b7ad9bc292b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.986250] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1060.986535] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdf2afc7-66fa-406b-aed9-fa8a49d77c78 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.068934] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.069197] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.069415] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleting the datastore file [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.069655] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc2798c4-eb9d-40e9-8213-b5996fc0e0dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.072662] env[65385]: DEBUG nova.scheduler.client.report [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1061.081388] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1061.081388] env[65385]: value = "task-4454315" [ 1061.081388] env[65385]: _type = "Task" [ 1061.081388] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.091736] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.325842] env[65385]: DEBUG nova.compute.utils [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1061.378433] env[65385]: DEBUG nova.compute.manager [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1061.378730] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.379707] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db18a5a3-fdb5-4ff6-becf-f0b56e9d514b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.388989] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.389394] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-521de850-b7b4-4f0d-ae45-d75aa467af06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.397839] env[65385]: DEBUG oslo_vmware.api [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1061.397839] env[65385]: value = "task-4454316" [ 1061.397839] env[65385]: _type = "Task" [ 1061.397839] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.407587] env[65385]: DEBUG oslo_vmware.api [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454316, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.430025] env[65385]: DEBUG nova.network.neutron [-] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1061.445181] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 60%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.544971] env[65385]: DEBUG nova.compute.manager [req-4b0f9476-68cf-40d9-b463-3c83a306b6af req-b418d285-b1e6-49dd-bd52-f6e161d46b55 service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-vif-deleted-14fd5163-1262-4edf-8c36-42b82ce49e2d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1061.578335] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.578702] env[65385]: DEBUG nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1061.582591] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.960s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.582886] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.585272] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.239s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.585469] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.585634] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1061.585954] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.685s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.587870] env[65385]: INFO nova.compute.claims [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1061.591930] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f206c2-1299-4cd8-bf5c-f4d56e383bfd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.608961] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991510e0-b306-4816-b7c1-f6657e133ff2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.613970] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.382152} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.614821] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.614821] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.614821] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.619392] env[65385]: INFO nova.scheduler.client.report [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Deleted allocations for instance 8a974e99-a7b1-438e-aace-9ae82352495c [ 1061.635485] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d4bab0-9c4c-4471-9751-8d3b1db81819 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.646033] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c071ca89-3ccc-4718-8b42-078dc5fa7141 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.686790] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179850MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1061.686790] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.829305] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.909983] env[65385]: DEBUG oslo_vmware.api [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454316, 'name': PowerOffVM_Task, 'duration_secs': 0.283064} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.910292] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.910455] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.910742] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf8ef803-dd3b-4664-8181-99910846383f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.932809] env[65385]: INFO nova.compute.manager [-] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Took 1.64 seconds to deallocate network for instance. [ 1061.949361] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 69%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.998172] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.998420] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.998578] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleting the datastore file [datastore2] 59f26752-75ab-45e5-b0a9-bbcef44ecd7f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.998909] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b85ffd9f-244a-45ee-9372-4c822aff8161 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.007850] env[65385]: DEBUG oslo_vmware.api [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1062.007850] env[65385]: value = "task-4454318" [ 1062.007850] env[65385]: _type = "Task" [ 1062.007850] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.019303] env[65385]: DEBUG oslo_vmware.api [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.084569] env[65385]: DEBUG nova.compute.utils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1062.085990] env[65385]: DEBUG nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1062.086123] env[65385]: DEBUG nova.network.neutron [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1062.086458] env[65385]: WARNING neutronclient.v2_0.client [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.087076] env[65385]: WARNING neutronclient.v2_0.client [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.087451] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.087873] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.138739] env[65385]: DEBUG oslo_concurrency.lockutils [None req-937f1a45-1c70-4897-80ac-8aadac003d8a tempest-DeleteServersTestJSON-564283273 tempest-DeleteServersTestJSON-564283273-project-member] Lock "8a974e99-a7b1-438e-aace-9ae82352495c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.999s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.153933] env[65385]: DEBUG nova.policy [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b450c500f84f2fa682d70698c398b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d45ff3446a84ef2a1945f7a66b747e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1062.443648] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.448680] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 82%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.523825] env[65385]: DEBUG oslo_vmware.api [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207316} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.524234] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.524403] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1062.524602] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1062.524794] env[65385]: INFO nova.compute.manager [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1062.525087] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1062.525348] env[65385]: DEBUG nova.compute.manager [-] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1062.525447] env[65385]: DEBUG nova.network.neutron [-] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1062.525724] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.526668] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.526962] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.535559] env[65385]: DEBUG nova.network.neutron [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Successfully created port: d33bca4a-a999-48ad-8138-b0f8d5b6ee23 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1062.595209] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.598007] env[65385]: DEBUG nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1062.679140] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1062.679717] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1062.679717] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1062.679717] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1062.679866] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1062.679958] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1062.680344] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.680543] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1062.680723] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1062.680879] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1062.681405] env[65385]: DEBUG nova.virt.hardware [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1062.682380] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1484054e-2934-4051-a163-2a0dc353c647 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.698672] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14353908-1575-4de2-bfb8-a8048cd73777 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.719566] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:1a:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0fc23e28-1265-46b7-aa43-fd72e2269330', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.727277] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1062.730432] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1062.731750] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bcf79e01-552a-4b53-bd0b-d77cdc59a350 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.757388] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.757388] env[65385]: value = "task-4454320" [ 1062.757388] env[65385]: _type = "Task" [ 1062.757388] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.769852] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454320, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.823436] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893a2ff7-b3cf-4ef8-8150-4ff8ff49369a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.836209] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ba83e2-2fdd-4e95-992e-13ba2476c99e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.873633] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26066429-1e9f-45e4-8d75-c8c9c7c6358c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.882147] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62eda87f-9617-44c9-a5c3-e37df6122e38 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.898506] env[65385]: DEBUG nova.compute.provider_tree [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.937233] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.937519] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.937839] env[65385]: INFO nova.compute.manager [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Attaching volume 9ecf961e-bf7d-4631-ad04-666b1111075d to /dev/sdb [ 1062.953497] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.983208] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03adf8cc-4ea2-4269-99d3-8bbdf8ae4d67 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.991411] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc3b644-1b52-4fdc-956d-5b4e7d51fbaa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.005889] env[65385]: DEBUG nova.virt.block_device [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating existing volume attachment record: a8a3e1a3-1452-4e34-bdf1-6b9ad709ead5 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1063.266869] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454320, 'name': CreateVM_Task, 'duration_secs': 0.444479} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.267115] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1063.267667] env[65385]: WARNING neutronclient.v2_0.client [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1063.268629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.268629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.268629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1063.268886] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aff853a-53b1-46a0-b651-4205248ebf9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.273489] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1063.273489] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d33518-cc13-4650-55cf-c53a9dbfef36" [ 1063.273489] env[65385]: _type = "Task" [ 1063.273489] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.282488] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d33518-cc13-4650-55cf-c53a9dbfef36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.402334] env[65385]: DEBUG nova.scheduler.client.report [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1063.455334] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.586017] env[65385]: DEBUG nova.compute.manager [req-e234a7ba-990f-41a9-bee8-a41ba23233f4 req-9e9c05a3-4377-44d0-bcd6-eab46aa7abe5 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Received event network-vif-deleted-59289962-8634-414f-9bd2-b3f5b19af9b4 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1063.586389] env[65385]: INFO nova.compute.manager [req-e234a7ba-990f-41a9-bee8-a41ba23233f4 req-9e9c05a3-4377-44d0-bcd6-eab46aa7abe5 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Neutron deleted interface 59289962-8634-414f-9bd2-b3f5b19af9b4; detaching it from the instance and deleting it from the info cache [ 1063.586667] env[65385]: DEBUG nova.network.neutron [req-e234a7ba-990f-41a9-bee8-a41ba23233f4 req-9e9c05a3-4377-44d0-bcd6-eab46aa7abe5 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.612077] env[65385]: DEBUG nova.network.neutron [-] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.614208] env[65385]: DEBUG nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1063.645175] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1063.645418] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1063.645592] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1063.645796] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1063.645962] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1063.646147] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1063.646407] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.646574] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1063.646763] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1063.646949] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1063.647158] env[65385]: DEBUG nova.virt.hardware [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1063.648089] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a414baf-1c55-4fe6-84bd-f1575bffeb94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.658836] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c651e2-a41e-4433-b4cd-16b1b68dff97 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.785540] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d33518-cc13-4650-55cf-c53a9dbfef36, 'name': SearchDatastore_Task, 'duration_secs': 0.011269} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.785876] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.786120] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.786359] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.786500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.786668] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.786935] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88946650-5b79-47e4-8474-9b898400e308 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.798079] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1063.798325] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1063.799092] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-323a2bee-10f8-48fa-be35-670bb8dc8f8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.806022] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1063.806022] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c985d-02b2-1acc-217d-6aac045f56e8" [ 1063.806022] env[65385]: _type = "Task" [ 1063.806022] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.815544] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c985d-02b2-1acc-217d-6aac045f56e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.908061] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.908729] env[65385]: DEBUG nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1063.911781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.225s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.952249] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.083327] env[65385]: DEBUG nova.compute.manager [req-03241809-11ba-4229-878c-ea1e986d9597 req-2c60f7f9-8461-4604-ba34-755700dd2e2d service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Received event network-vif-plugged-d33bca4a-a999-48ad-8138-b0f8d5b6ee23 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1064.083603] env[65385]: DEBUG oslo_concurrency.lockutils [req-03241809-11ba-4229-878c-ea1e986d9597 req-2c60f7f9-8461-4604-ba34-755700dd2e2d service nova] Acquiring lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.083747] env[65385]: DEBUG oslo_concurrency.lockutils [req-03241809-11ba-4229-878c-ea1e986d9597 req-2c60f7f9-8461-4604-ba34-755700dd2e2d service nova] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.083922] env[65385]: DEBUG oslo_concurrency.lockutils [req-03241809-11ba-4229-878c-ea1e986d9597 req-2c60f7f9-8461-4604-ba34-755700dd2e2d service nova] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1064.084123] env[65385]: DEBUG nova.compute.manager [req-03241809-11ba-4229-878c-ea1e986d9597 req-2c60f7f9-8461-4604-ba34-755700dd2e2d service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] No waiting events found dispatching network-vif-plugged-d33bca4a-a999-48ad-8138-b0f8d5b6ee23 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1064.084294] env[65385]: WARNING nova.compute.manager [req-03241809-11ba-4229-878c-ea1e986d9597 req-2c60f7f9-8461-4604-ba34-755700dd2e2d service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Received unexpected event network-vif-plugged-d33bca4a-a999-48ad-8138-b0f8d5b6ee23 for instance with vm_state building and task_state spawning. [ 1064.089493] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1088db42-8ebf-4d98-9328-a55e64f9b4a5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.099019] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6df87f-bdfa-4f56-9482-85b4d022e94c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.117868] env[65385]: INFO nova.compute.manager [-] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Took 1.59 seconds to deallocate network for instance. [ 1064.131772] env[65385]: DEBUG nova.compute.manager [req-e234a7ba-990f-41a9-bee8-a41ba23233f4 req-9e9c05a3-4377-44d0-bcd6-eab46aa7abe5 service nova] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Detach interface failed, port_id=59289962-8634-414f-9bd2-b3f5b19af9b4, reason: Instance 59f26752-75ab-45e5-b0a9-bbcef44ecd7f could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1064.172862] env[65385]: DEBUG nova.network.neutron [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Successfully updated port: d33bca4a-a999-48ad-8138-b0f8d5b6ee23 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1064.317721] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526c985d-02b2-1acc-217d-6aac045f56e8, 'name': SearchDatastore_Task, 'duration_secs': 0.010625} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.318705] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-356247ef-80e1-4eff-a045-4086271f6b34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.326365] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1064.326365] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261943d-ba14-0a5c-faa4-bc31853f131e" [ 1064.326365] env[65385]: _type = "Task" [ 1064.326365] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.335536] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261943d-ba14-0a5c-faa4-bc31853f131e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.416045] env[65385]: DEBUG nova.compute.utils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1064.422695] env[65385]: DEBUG nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1064.422903] env[65385]: DEBUG nova.network.neutron [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1064.424339] env[65385]: WARNING neutronclient.v2_0.client [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.424339] env[65385]: WARNING neutronclient.v2_0.client [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.424339] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.424568] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.453559] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.469997] env[65385]: DEBUG nova.policy [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662871f375e441eb98d814cf7d7ac5c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bc8afc699e34f059f18d8244980bc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1064.634469] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.675735] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.675914] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.676175] env[65385]: DEBUG nova.network.neutron [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1064.842057] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5261943d-ba14-0a5c-faa4-bc31853f131e, 'name': SearchDatastore_Task, 'duration_secs': 0.010988} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.843542] env[65385]: DEBUG nova.network.neutron [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Successfully created port: 32b4dfff-31d2-41e4-9509-668e6d7febf5 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1064.847063] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1064.847377] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1064.848023] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1949729b-1a5e-4d5a-bafe-f0473938c51b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.858468] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1064.858468] env[65385]: value = "task-4454324" [ 1064.858468] env[65385]: _type = "Task" [ 1064.858468] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.869445] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454324, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.923603] env[65385]: DEBUG nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1064.956120] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task} progress is 98%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.968640] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.968817] env[65385]: WARNING nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance a5fe2ad9-6dfe-485b-bf27-beb218d31467 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1064.968940] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969084] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b4bca2c8-fc8f-471a-899b-aac648fee9fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969233] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 5470afe7-6dd0-4778-9b83-2956ede04b7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969348] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 992920c3-af4f-489e-b21f-9d52c24399de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969461] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 59f26752-75ab-45e5-b0a9-bbcef44ecd7f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969723] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 77d718c8-4959-483a-8717-c00f2be1d0bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969723] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969806] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.969859] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 04ebe83c-cab5-45e1-9f5f-018481f63912 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1064.970070] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1064.970225] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=100GB used_disk=9GB total_vcpus=48 used_vcpus=10 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '10', 'num_vm_active': '7', 'num_task_deleting': '1', 'num_os_type_None': '10', 'num_proj_60de0d1c162342209795a180391661f6': '1', 'io_workload': '3', 'num_task_None': '6', 'num_proj_b27c092a0fd345ea953956469c21f8bb': '1', 'num_proj_1598307c304d46a884a88b3eb5f5d7dc': '1', 'num_proj_8d45ff3446a84ef2a1945f7a66b747e8': '2', 'num_proj_2ba61d77cc2d4fe1ba3c03466d5985d9': '1', 'num_proj_6bc8afc699e34f059f18d8244980bc1d': '2', 'num_task_rebuild_spawning': '1', 'num_proj_94cc5bc221c0455ea760f5022db6bdc4': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_87efee8d30ee4a69bdd8dc784e809bfe': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1065.151815] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed883282-b355-432e-a80a-0fe6ee462d79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.163987] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934b4310-3e38-46b2-af57-a9cd54b915aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.200557] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.201449] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.212996] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f2ec77-b71e-45f8-8bc8-9061ce25d84b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.228788] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41cce83-21da-4b53-af29-013eae05fc57 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.251250] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.287175] env[65385]: DEBUG nova.network.neutron [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1065.336733] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.337201] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.372126] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454324, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.456194] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454312, 'name': RelocateVM_Task, 'duration_secs': 5.064445} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.457821] env[65385]: WARNING neutronclient.v2_0.client [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.458761] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.459654] env[65385]: WARNING openstack [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.469145] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1065.469440] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871157', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'name': 'volume-aee71a50-6144-4280-9b67-ff58bba61633', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b1abf6ab-a3d5-4078-9568-68db0a7f03cc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'serial': 'aee71a50-6144-4280-9b67-ff58bba61633'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1065.470779] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae07a96c-dffb-4775-ad97-4948a554c293 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.493473] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498093b5-3369-433d-81f4-d31ba88e59a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.524083] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-aee71a50-6144-4280-9b67-ff58bba61633/volume-aee71a50-6144-4280-9b67-ff58bba61633.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.524524] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e1a8988-3099-4802-a482-ae13c0a42fd8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.550278] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1065.550278] env[65385]: value = "task-4454326" [ 1065.550278] env[65385]: _type = "Task" [ 1065.550278] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.565344] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454326, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.670801] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.671100] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.755975] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1065.871863] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454324, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.760496} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.872108] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1065.872196] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1065.872467] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7cebcdfc-947b-4e25-bc3a-9e4d3fe029d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.882909] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1065.882909] env[65385]: value = "task-4454327" [ 1065.882909] env[65385]: _type = "Task" [ 1065.882909] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.897948] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.936196] env[65385]: DEBUG nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1065.969736] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1065.970119] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1065.970290] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1065.970509] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1065.970649] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1065.970791] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1065.971019] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1065.971197] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1065.971576] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1065.971669] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1065.971765] env[65385]: DEBUG nova.virt.hardware [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1065.972993] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52cc46f-3b7d-4784-8722-3af2dc264567 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.983292] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c121bcd6-d0db-4215-80c1-bb36ceceeada {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.061074] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454326, 'name': ReconfigVM_Task, 'duration_secs': 0.461845} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.061406] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-aee71a50-6144-4280-9b67-ff58bba61633/volume-aee71a50-6144-4280-9b67-ff58bba61633.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.068167] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90377dad-7414-418e-8968-6a3cf6e5cc36 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.092939] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1066.092939] env[65385]: value = "task-4454328" [ 1066.092939] env[65385]: _type = "Task" [ 1066.092939] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.105960] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454328, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.174426] env[65385]: INFO nova.compute.manager [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Detaching volume fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2 [ 1066.228024] env[65385]: INFO nova.virt.block_device [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Attempting to driver detach volume fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2 from mountpoint /dev/sdb [ 1066.228024] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1066.228024] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871137', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'name': 'volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'serial': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1066.229197] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3b73a7-580c-4def-b669-08626ab07a03 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.253082] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279b097e-142c-487f-bbf5-cdf481d29155 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.263451] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1066.263451] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.351s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.263688] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.820s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.264121] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.267068] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.633s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.267489] env[65385]: DEBUG nova.objects.instance [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'resources' on Instance uuid 59f26752-75ab-45e5-b0a9-bbcef44ecd7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.272730] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1667e7-0c9e-40d9-9782-0284dfba56b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.301794] env[65385]: INFO nova.scheduler.client.report [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted allocations for instance a5fe2ad9-6dfe-485b-bf27-beb218d31467 [ 1066.303361] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15500b6c-7fb3-4aa0-9d90-6d2d5cce59ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.327064] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The volume has not been displaced from its original location: [datastore1] volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2/volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1066.332748] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1066.336390] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35befe8d-9740-4e20-978a-8b35c098076c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.354604] env[65385]: DEBUG oslo_vmware.api [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1066.354604] env[65385]: value = "task-4454329" [ 1066.354604] env[65385]: _type = "Task" [ 1066.354604] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.371341] env[65385]: DEBUG oslo_vmware.api [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454329, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.395731] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454327, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081881} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.395731] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.396377] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b4181d7-141d-4ad6-ad5c-31067a27e993 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.421423] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.421892] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab56d5a9-c80f-4d83-8726-8f7951debeda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.447680] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1066.447680] env[65385]: value = "task-4454330" [ 1066.447680] env[65385]: _type = "Task" [ 1066.447680] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.451679] env[65385]: DEBUG nova.network.neutron [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Successfully updated port: 32b4dfff-31d2-41e4-9509-668e6d7febf5 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1066.459727] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454330, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.604053] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454328, 'name': ReconfigVM_Task, 'duration_secs': 0.419857} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.604447] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871157', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'name': 'volume-aee71a50-6144-4280-9b67-ff58bba61633', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b1abf6ab-a3d5-4078-9568-68db0a7f03cc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'serial': 'aee71a50-6144-4280-9b67-ff58bba61633'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1066.605151] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce7fd099-8069-4d00-a506-c02087a0609f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.613397] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1066.613397] env[65385]: value = "task-4454331" [ 1066.613397] env[65385]: _type = "Task" [ 1066.613397] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.615622] env[65385]: DEBUG nova.network.neutron [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Updating instance_info_cache with network_info: [{"id": "d33bca4a-a999-48ad-8138-b0f8d5b6ee23", "address": "fa:16:3e:ab:73:20", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd33bca4a-a9", "ovs_interfaceid": "d33bca4a-a999-48ad-8138-b0f8d5b6ee23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1066.626596] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454331, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.727305] env[65385]: DEBUG nova.compute.manager [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Received event network-changed-d33bca4a-a999-48ad-8138-b0f8d5b6ee23 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1066.727510] env[65385]: DEBUG nova.compute.manager [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Refreshing instance network info cache due to event network-changed-d33bca4a-a999-48ad-8138-b0f8d5b6ee23. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1066.727697] env[65385]: DEBUG oslo_concurrency.lockutils [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Acquiring lock "refresh_cache-5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.784049] env[65385]: DEBUG nova.compute.manager [req-19c8b7c9-c22f-447a-9d65-fcd643cac8fb req-56323a29-7b57-436c-9050-87dabf620dc2 service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Received event network-vif-plugged-32b4dfff-31d2-41e4-9509-668e6d7febf5 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1066.784179] env[65385]: DEBUG oslo_concurrency.lockutils [req-19c8b7c9-c22f-447a-9d65-fcd643cac8fb req-56323a29-7b57-436c-9050-87dabf620dc2 service nova] Acquiring lock "04ebe83c-cab5-45e1-9f5f-018481f63912-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.784752] env[65385]: DEBUG oslo_concurrency.lockutils [req-19c8b7c9-c22f-447a-9d65-fcd643cac8fb req-56323a29-7b57-436c-9050-87dabf620dc2 service nova] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.785198] env[65385]: DEBUG oslo_concurrency.lockutils [req-19c8b7c9-c22f-447a-9d65-fcd643cac8fb req-56323a29-7b57-436c-9050-87dabf620dc2 service nova] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.785198] env[65385]: DEBUG nova.compute.manager [req-19c8b7c9-c22f-447a-9d65-fcd643cac8fb req-56323a29-7b57-436c-9050-87dabf620dc2 service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] No waiting events found dispatching network-vif-plugged-32b4dfff-31d2-41e4-9509-668e6d7febf5 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1066.785344] env[65385]: WARNING nova.compute.manager [req-19c8b7c9-c22f-447a-9d65-fcd643cac8fb req-56323a29-7b57-436c-9050-87dabf620dc2 service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Received unexpected event network-vif-plugged-32b4dfff-31d2-41e4-9509-668e6d7febf5 for instance with vm_state building and task_state spawning. [ 1066.819730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8cb6ac7-a0d1-4b04-9f97-c6d8ec7ea60f tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.208s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.820726] env[65385]: DEBUG oslo_concurrency.lockutils [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] Acquired lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.821760] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb16e26-128a-4597-9b25-4bcc5f2d2981 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.832995] env[65385]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1066.833203] env[65385]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=65385) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1066.833947] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5291075c-d103-44e5-8b80-7b7f0503c440 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.847722] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dce2025-d35f-48bd-9d8c-1986a6d586b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.872923] env[65385]: DEBUG oslo_vmware.api [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454329, 'name': ReconfigVM_Task, 'duration_secs': 0.426415} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.875708] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1066.890235] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ad9963f-d8d6-40e5-a363-9948c73a7c91 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.901771] env[65385]: ERROR root [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-871133' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 479, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-871133' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-871133' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-871133'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-871133' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-871133' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-871133'}\n"]: nova.exception.InstanceNotFound: Instance a5fe2ad9-6dfe-485b-bf27-beb218d31467 could not be found. [ 1066.902124] env[65385]: DEBUG oslo_concurrency.lockutils [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] Releasing lock "a5fe2ad9-6dfe-485b-bf27-beb218d31467" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.902238] env[65385]: DEBUG nova.compute.manager [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Detach interface failed, port_id=ef6b6d8f-dd4a-4682-991b-9c6dd0d27525, reason: Instance a5fe2ad9-6dfe-485b-bf27-beb218d31467 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1066.902379] env[65385]: DEBUG nova.compute.manager [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Received event network-vif-deleted-05d2fd35-d0a1-44d7-9944-0b6445e2bd9b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1066.902545] env[65385]: INFO nova.compute.manager [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Neutron deleted interface 05d2fd35-d0a1-44d7-9944-0b6445e2bd9b; detaching it from the instance and deleting it from the info cache [ 1066.902805] env[65385]: DEBUG nova.network.neutron [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Updating instance_info_cache with network_info: [{"id": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "address": "fa:16:3e:52:d2:85", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14fd5163-12", "ovs_interfaceid": "14fd5163-1262-4edf-8c36-42b82ce49e2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1066.913785] env[65385]: DEBUG oslo_vmware.api [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1066.913785] env[65385]: value = "task-4454332" [ 1066.913785] env[65385]: _type = "Task" [ 1066.913785] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.929612] env[65385]: DEBUG oslo_vmware.api [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454332, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.959165] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "refresh_cache-04ebe83c-cab5-45e1-9f5f-018481f63912" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.959550] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "refresh_cache-04ebe83c-cab5-45e1-9f5f-018481f63912" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.960195] env[65385]: DEBUG nova.network.neutron [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1066.961880] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454330, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.981986] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a8a16d-f870-4fa4-ba63-120a6235a9fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.992358] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e78d11-9abf-4fda-a246-01879850ee8e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.026563] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b9ed8e-3465-4a44-8950-1ff8f181bc03 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.036897] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57114f68-e464-47bc-a279-0f42f6c30e0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.056489] env[65385]: DEBUG nova.compute.provider_tree [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.122563] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1067.122990] env[65385]: DEBUG nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Instance network_info: |[{"id": "d33bca4a-a999-48ad-8138-b0f8d5b6ee23", "address": "fa:16:3e:ab:73:20", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd33bca4a-a9", "ovs_interfaceid": "d33bca4a-a999-48ad-8138-b0f8d5b6ee23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1067.124460] env[65385]: DEBUG oslo_concurrency.lockutils [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Acquired lock "refresh_cache-5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.124460] env[65385]: DEBUG nova.network.neutron [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Refreshing network info cache for port d33bca4a-a999-48ad-8138-b0f8d5b6ee23 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1067.125524] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:73:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd33bca4a-a999-48ad-8138-b0f8d5b6ee23', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.134281] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1067.135541] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.135799] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-252706fa-a5b1-4216-b43c-a0a9a8b2d2c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.158146] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454331, 'name': Rename_Task, 'duration_secs': 0.146658} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.158274] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.160684] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a02da5a-c227-49eb-a348-d56870dac648 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.164292] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.164292] env[65385]: value = "task-4454333" [ 1067.164292] env[65385]: _type = "Task" [ 1067.164292] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.170477] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1067.170477] env[65385]: value = "task-4454334" [ 1067.170477] env[65385]: _type = "Task" [ 1067.170477] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.178202] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454333, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.185930] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.263835] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.263835] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.263835] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.263835] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.408099] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-732e82bd-7eb0-4b2e-9905-1d4171d19975 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.421356] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729c61b9-9c6d-4fc5-b540-72feef0fe745 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.443075] env[65385]: DEBUG oslo_vmware.api [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454332, 'name': ReconfigVM_Task, 'duration_secs': 0.16953} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.443410] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871137', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'name': 'volume-fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2', 'serial': 'fb6cba7f-ffb0-4fef-9fef-8a786f69f2a2'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1067.456838] env[65385]: DEBUG nova.compute.manager [req-51e285a3-7eda-476b-87ab-58de124ae31a req-47ddcf4f-d7dd-4809-b3a9-1e6cbbaf175c service nova] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Detach interface failed, port_id=05d2fd35-d0a1-44d7-9944-0b6445e2bd9b, reason: Instance a5fe2ad9-6dfe-485b-bf27-beb218d31467 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1067.463809] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.464716] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.475862] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454330, 'name': ReconfigVM_Task, 'duration_secs': 0.838524} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.476318] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb/77d718c8-4959-483a-8717-c00f2be1d0bb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.476980] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2810d948-9358-4569-8f06-5ed93e5b3b5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.485090] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1067.485090] env[65385]: value = "task-4454335" [ 1067.485090] env[65385]: _type = "Task" [ 1067.485090] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.495452] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454335, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.560332] env[65385]: DEBUG nova.scheduler.client.report [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1067.598179] env[65385]: DEBUG nova.network.neutron [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1067.636583] env[65385]: WARNING neutronclient.v2_0.client [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.637872] env[65385]: WARNING openstack [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.638040] env[65385]: WARNING openstack [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.663081] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.663468] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.690053] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454333, 'name': CreateVM_Task, 'duration_secs': 0.384006} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.693893] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.694392] env[65385]: DEBUG oslo_vmware.api [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454334, 'name': PowerOnVM_Task, 'duration_secs': 0.504858} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.694893] env[65385]: WARNING neutronclient.v2_0.client [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.696079] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.696079] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.696079] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1067.697496] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.697608] env[65385]: INFO nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Took 8.74 seconds to spawn the instance on the hypervisor. [ 1067.697790] env[65385]: DEBUG nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1067.698088] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f67ada8-4143-4ff7-9e27-4df7a0b133c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.700713] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5fcce6d-b7ce-4cb3-908d-1c0d34b2043c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.715222] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1067.715222] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a761fe-9def-2bfd-6132-5767e747a29a" [ 1067.715222] env[65385]: _type = "Task" [ 1067.715222] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.726331] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a761fe-9def-2bfd-6132-5767e747a29a, 'name': SearchDatastore_Task, 'duration_secs': 0.013473} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.727035] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1067.727035] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.727349] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.727508] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.727685] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.728158] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7eba7a9-90cd-44d4-9e85-5e36be9087c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.745467] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.745664] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.746527] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c4b744b-f81f-4e6d-96e9-8a23ea4bea1c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.754163] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1067.754163] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5273a7a4-e027-2fa2-b7ad-4dae69836ab9" [ 1067.754163] env[65385]: _type = "Task" [ 1067.754163] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.764678] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5273a7a4-e027-2fa2-b7ad-4dae69836ab9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.801629] env[65385]: WARNING neutronclient.v2_0.client [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.802516] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.802904] env[65385]: WARNING openstack [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.926065] env[65385]: WARNING openstack [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.926526] env[65385]: WARNING openstack [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.996929] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454335, 'name': Rename_Task, 'duration_secs': 0.167378} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.998123] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.998123] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7068a0a0-7b6c-4d4f-9f7c-adc9f029fe65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.008457] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1068.008457] env[65385]: value = "task-4454336" [ 1068.008457] env[65385]: _type = "Task" [ 1068.008457] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.011728] env[65385]: DEBUG nova.objects.instance [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'flavor' on Instance uuid 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.020478] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.066396] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1068.066937] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871164', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'name': 'volume-9ecf961e-bf7d-4631-ad04-666b1111075d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '29a449d0-f62d-4430-96ab-9ac43e5a7ad2', 'attached_at': '', 'detached_at': '', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'serial': '9ecf961e-bf7d-4631-ad04-666b1111075d'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1068.068849] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.802s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.074179] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564ab85f-588d-4b3d-a209-29c8a8e3927d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.098081] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbb3cd7-7412-467a-981d-1bf80e134132 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.137306] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] volume-9ecf961e-bf7d-4631-ad04-666b1111075d/volume-9ecf961e-bf7d-4631-ad04-666b1111075d.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.138550] env[65385]: INFO nova.scheduler.client.report [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted allocations for instance 59f26752-75ab-45e5-b0a9-bbcef44ecd7f [ 1068.142198] env[65385]: DEBUG nova.network.neutron [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Updating instance_info_cache with network_info: [{"id": "32b4dfff-31d2-41e4-9509-668e6d7febf5", "address": "fa:16:3e:1c:af:4c", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32b4dfff-31", "ovs_interfaceid": "32b4dfff-31d2-41e4-9509-668e6d7febf5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1068.142872] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6173311-08e5-4f7a-8fb6-b89e1d8d78e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.159070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "refresh_cache-04ebe83c-cab5-45e1-9f5f-018481f63912" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.159437] env[65385]: DEBUG nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Instance network_info: |[{"id": "32b4dfff-31d2-41e4-9509-668e6d7febf5", "address": "fa:16:3e:1c:af:4c", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32b4dfff-31", "ovs_interfaceid": "32b4dfff-31d2-41e4-9509-668e6d7febf5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1068.163115] env[65385]: WARNING neutronclient.v2_0.client [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.163781] env[65385]: WARNING openstack [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.164226] env[65385]: WARNING openstack [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.172316] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:af:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32b4dfff-31d2-41e4-9509-668e6d7febf5', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1068.179869] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1068.181375] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1068.182300] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0004f400-6fae-4b24-ad41-5786bbc473a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.204366] env[65385]: DEBUG oslo_vmware.api [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1068.204366] env[65385]: value = "task-4454337" [ 1068.204366] env[65385]: _type = "Task" [ 1068.204366] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.213198] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1068.213198] env[65385]: value = "task-4454338" [ 1068.213198] env[65385]: _type = "Task" [ 1068.213198] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.226414] env[65385]: DEBUG oslo_vmware.api [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.229546] env[65385]: INFO nova.compute.manager [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Took 15.97 seconds to build instance. [ 1068.239703] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454338, 'name': CreateVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.266934] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5273a7a4-e027-2fa2-b7ad-4dae69836ab9, 'name': SearchDatastore_Task, 'duration_secs': 0.014292} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.268305] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ce82006-2987-403b-a4a6-0c16044e40ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.275899] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1068.275899] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522a2d94-1261-06a9-ec1e-9ac7503d7570" [ 1068.275899] env[65385]: _type = "Task" [ 1068.275899] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.286744] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522a2d94-1261-06a9-ec1e-9ac7503d7570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.319253] env[65385]: DEBUG nova.network.neutron [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Updated VIF entry in instance network info cache for port d33bca4a-a999-48ad-8138-b0f8d5b6ee23. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1068.319656] env[65385]: DEBUG nova.network.neutron [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Updating instance_info_cache with network_info: [{"id": "d33bca4a-a999-48ad-8138-b0f8d5b6ee23", "address": "fa:16:3e:ab:73:20", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd33bca4a-a9", "ovs_interfaceid": "d33bca4a-a999-48ad-8138-b0f8d5b6ee23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1068.521341] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454336, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.684363] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7387389b-8e60-463c-943c-ee83a9834065 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "59f26752-75ab-45e5-b0a9-bbcef44ecd7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.814s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.719303] env[65385]: DEBUG oslo_vmware.api [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454337, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.732643] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f530a0fa-d4cf-46c1-ab2b-8e72d63bf027 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.482s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.733135] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454338, 'name': CreateVM_Task, 'duration_secs': 0.44118} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.733825] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1068.734487] env[65385]: WARNING neutronclient.v2_0.client [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.734941] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.735190] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.735534] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1068.735890] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ace5c589-8302-4795-b168-202eb78d6d44 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.744638] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1068.744638] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d1bab4-4ce6-0663-348b-8f4624dd53e0" [ 1068.744638] env[65385]: _type = "Task" [ 1068.744638] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.754795] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d1bab4-4ce6-0663-348b-8f4624dd53e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.787698] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522a2d94-1261-06a9-ec1e-9ac7503d7570, 'name': SearchDatastore_Task, 'duration_secs': 0.022381} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.787976] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.789173] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.789173] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97532d87-2940-490f-bf0c-583e129502ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.798485] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1068.798485] env[65385]: value = "task-4454339" [ 1068.798485] env[65385]: _type = "Task" [ 1068.798485] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.809753] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.822814] env[65385]: DEBUG oslo_concurrency.lockutils [req-1973fb08-9a99-436f-896b-d64de29eb600 req-7c78c4c9-fa37-4a8f-bf1b-12be85592fb5 service nova] Releasing lock "refresh_cache-5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1069.018307] env[65385]: DEBUG oslo_vmware.api [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454336, 'name': PowerOnVM_Task, 'duration_secs': 0.698729} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.018692] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1069.018809] env[65385]: DEBUG nova.compute.manager [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1069.019644] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1480ba3f-0bce-4074-a0f0-abfd7b3ea4fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.029451] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e75e8df9-24f5-457c-a2b6-e3f35424bf7b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.358s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.186604] env[65385]: DEBUG nova.compute.manager [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Received event network-changed-32b4dfff-31d2-41e4-9509-668e6d7febf5 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1069.186818] env[65385]: DEBUG nova.compute.manager [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Refreshing instance network info cache due to event network-changed-32b4dfff-31d2-41e4-9509-668e6d7febf5. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1069.187161] env[65385]: DEBUG oslo_concurrency.lockutils [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Acquiring lock "refresh_cache-04ebe83c-cab5-45e1-9f5f-018481f63912" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.187357] env[65385]: DEBUG oslo_concurrency.lockutils [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Acquired lock "refresh_cache-04ebe83c-cab5-45e1-9f5f-018481f63912" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1069.187357] env[65385]: DEBUG nova.network.neutron [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Refreshing network info cache for port 32b4dfff-31d2-41e4-9509-668e6d7febf5 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1069.216440] env[65385]: DEBUG oslo_vmware.api [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454337, 'name': ReconfigVM_Task, 'duration_secs': 0.54136} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.216823] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfigured VM instance instance-00000050 to attach disk [datastore1] volume-9ecf961e-bf7d-4631-ad04-666b1111075d/volume-9ecf961e-bf7d-4631-ad04-666b1111075d.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.221743] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6e1e36f-a889-4e08-872b-66e26c9fda8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.243300] env[65385]: DEBUG oslo_vmware.api [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1069.243300] env[65385]: value = "task-4454340" [ 1069.243300] env[65385]: _type = "Task" [ 1069.243300] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.262041] env[65385]: DEBUG oslo_vmware.api [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454340, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.266383] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d1bab4-4ce6-0663-348b-8f4624dd53e0, 'name': SearchDatastore_Task, 'duration_secs': 0.023304} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.266762] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1069.267064] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1069.267439] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.267562] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1069.267797] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1069.268520] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5deee0f-7558-4a1b-9a5f-2ae17cc4e16b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.280460] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1069.280673] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1069.281577] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-590cce34-95ef-4fe8-b0d7-2f98a4d3fa0f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.289626] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1069.289626] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f0a7a-ed3b-f8e7-9a39-7286c8227f10" [ 1069.289626] env[65385]: _type = "Task" [ 1069.289626] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.304354] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f0a7a-ed3b-f8e7-9a39-7286c8227f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.317266] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454339, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.545990] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.546395] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.547415] env[65385]: DEBUG nova.objects.instance [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1069.690602] env[65385]: WARNING neutronclient.v2_0.client [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1069.691462] env[65385]: WARNING openstack [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.691821] env[65385]: WARNING openstack [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.757896] env[65385]: DEBUG oslo_vmware.api [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454340, 'name': ReconfigVM_Task, 'duration_secs': 0.197208} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.757896] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871164', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'name': 'volume-9ecf961e-bf7d-4631-ad04-666b1111075d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '29a449d0-f62d-4430-96ab-9ac43e5a7ad2', 'attached_at': '', 'detached_at': '', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'serial': '9ecf961e-bf7d-4631-ad04-666b1111075d'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1069.802265] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520f0a7a-ed3b-f8e7-9a39-7286c8227f10, 'name': SearchDatastore_Task, 'duration_secs': 0.028254} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.804531] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e677c1c3-46a0-474e-9212-2d6e07f302a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.829026] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1069.829026] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c086bc-bf8f-f0c2-6a4e-f9497cec0fde" [ 1069.829026] env[65385]: _type = "Task" [ 1069.829026] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.829804] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.853343} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.830362] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.830488] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1069.835568] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fda9d5a1-e9b3-4fb8-a25c-531d55888722 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.852144] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1069.852144] env[65385]: value = "task-4454341" [ 1069.852144] env[65385]: _type = "Task" [ 1069.852144] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.852407] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c086bc-bf8f-f0c2-6a4e-f9497cec0fde, 'name': SearchDatastore_Task, 'duration_secs': 0.020867} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.854083] env[65385]: WARNING openstack [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.854531] env[65385]: WARNING openstack [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.861741] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1069.862268] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 04ebe83c-cab5-45e1-9f5f-018481f63912/04ebe83c-cab5-45e1-9f5f-018481f63912.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1069.870556] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68ff4b77-e9a4-47a1-8065-6c58f8408679 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.875627] env[65385]: DEBUG oslo_concurrency.lockutils [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.875895] env[65385]: DEBUG oslo_concurrency.lockutils [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.883336] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454341, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.886228] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1069.886228] env[65385]: value = "task-4454342" [ 1069.886228] env[65385]: _type = "Task" [ 1069.886228] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.908176] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454342, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.960269] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.961288] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "900639b6-9b98-436f-aaad-ea8391f67393" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.008349] env[65385]: WARNING neutronclient.v2_0.client [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.009368] env[65385]: WARNING openstack [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.009963] env[65385]: WARNING openstack [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.114878] env[65385]: DEBUG nova.network.neutron [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Updated VIF entry in instance network info cache for port 32b4dfff-31d2-41e4-9509-668e6d7febf5. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1070.115281] env[65385]: DEBUG nova.network.neutron [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Updating instance_info_cache with network_info: [{"id": "32b4dfff-31d2-41e4-9509-668e6d7febf5", "address": "fa:16:3e:1c:af:4c", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32b4dfff-31", "ovs_interfaceid": "32b4dfff-31d2-41e4-9509-668e6d7febf5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1070.245538] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.245870] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.246182] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.246449] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.246630] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.249124] env[65385]: INFO nova.compute.manager [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Terminating instance [ 1070.362896] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "77d718c8-4959-483a-8717-c00f2be1d0bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.363166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.363384] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "77d718c8-4959-483a-8717-c00f2be1d0bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.363615] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.363837] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.369867] env[65385]: INFO nova.compute.manager [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Terminating instance [ 1070.378264] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454341, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.386350] env[65385]: DEBUG nova.compute.utils [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1070.399234] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454342, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.462883] env[65385]: DEBUG nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1070.555907] env[65385]: DEBUG oslo_concurrency.lockutils [None req-15d36c0e-f45b-41e7-b660-2750e45541d8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.617880] env[65385]: DEBUG oslo_concurrency.lockutils [req-c668f05d-83f3-4f9e-a1a3-26e86cadceda req-9fe10feb-cd73-43b2-b18f-85279bcbf5ca service nova] Releasing lock "refresh_cache-04ebe83c-cab5-45e1-9f5f-018481f63912" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.754013] env[65385]: DEBUG nova.compute.manager [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1070.754289] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.755248] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985c4532-badf-4634-8547-38f0d5b60a2c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.764244] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.764532] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dab25fbf-99aa-411b-8836-03a51d803884 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.771881] env[65385]: DEBUG oslo_vmware.api [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1070.771881] env[65385]: value = "task-4454343" [ 1070.771881] env[65385]: _type = "Task" [ 1070.771881] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.782712] env[65385]: DEBUG oslo_vmware.api [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454343, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.803148] env[65385]: DEBUG nova.objects.instance [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'flavor' on Instance uuid 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.876985] env[65385]: DEBUG nova.compute.manager [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1070.877241] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.877507] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454341, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.006024} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.878398] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b324c7-0441-4fb6-8890-e37839c14d9f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.882140] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1070.882867] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d489ab5-e7b1-4645-9cec-1109ec108e8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.905267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.029s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.905865] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.915710] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.916599] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dd763d1-ea06-4ce7-bb76-29fda21f0d5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.918432] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e83bda1-762b-45ee-b471-4c80cab914d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.942766] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454342, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.950725] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1070.950725] env[65385]: value = "task-4454345" [ 1070.950725] env[65385]: _type = "Task" [ 1070.950725] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.951988] env[65385]: DEBUG oslo_vmware.api [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1070.951988] env[65385]: value = "task-4454344" [ 1070.951988] env[65385]: _type = "Task" [ 1070.951988] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.979262] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454345, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.979582] env[65385]: DEBUG oslo_vmware.api [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454344, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.997934] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.998335] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.000243] env[65385]: INFO nova.compute.claims [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.144304] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "6c826914-4c6b-456b-a403-2373d30c803f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.144672] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.286244] env[65385]: DEBUG oslo_vmware.api [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454343, 'name': PowerOffVM_Task, 'duration_secs': 0.249469} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.286700] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.286981] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1071.287419] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7e5009c-e8fa-46c2-bcb2-bac8cc5b90fd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.309313] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebb71dbb-751c-41d8-b3ff-d93a895ec2ec tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.372s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1071.399247] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454342, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.429025} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.399576] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 04ebe83c-cab5-45e1-9f5f-018481f63912/04ebe83c-cab5-45e1-9f5f-018481f63912.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1071.400330] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.400407] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5159ae8-4fcc-4a63-ba4d-93d63ed07f58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.408484] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1071.408484] env[65385]: value = "task-4454347" [ 1071.408484] env[65385]: _type = "Task" [ 1071.408484] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.419496] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454347, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.470734] env[65385]: DEBUG oslo_vmware.api [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454344, 'name': PowerOffVM_Task, 'duration_secs': 0.235989} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.470981] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454345, 'name': ReconfigVM_Task, 'duration_secs': 0.428076} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.472123] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.472123] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1071.472123] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.472795] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-daffd677-1bc0-4da3-bf6a-64279eed8b23 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.476024] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35aded43-f95a-4339-999f-c252c6e4f676 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.485056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1071.485393] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1071.485490] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleting the datastore file [datastore1] 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.487140] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb638666-4f7f-4f7f-afce-e73e69881495 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.489707] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1071.489707] env[65385]: value = "task-4454349" [ 1071.489707] env[65385]: _type = "Task" [ 1071.489707] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.495912] env[65385]: DEBUG oslo_vmware.api [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1071.495912] env[65385]: value = "task-4454350" [ 1071.495912] env[65385]: _type = "Task" [ 1071.495912] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.502902] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454349, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.512082] env[65385]: DEBUG oslo_vmware.api [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.552133] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "805528b8-96db-4bb8-ba05-2b6332c54624" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.552349] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.556741] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1071.557032] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1071.557294] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleting the datastore file [datastore1] 77d718c8-4959-483a-8717-c00f2be1d0bb {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.558240] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ebecdf5-10ad-4492-88a8-c8a32f52fa51 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.566015] env[65385]: DEBUG oslo_vmware.api [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1071.566015] env[65385]: value = "task-4454351" [ 1071.566015] env[65385]: _type = "Task" [ 1071.566015] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.577463] env[65385]: DEBUG oslo_vmware.api [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454351, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.647282] env[65385]: DEBUG nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1071.651020] env[65385]: DEBUG nova.compute.manager [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Received event network-changed-b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1071.651240] env[65385]: DEBUG nova.compute.manager [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Refreshing instance network info cache due to event network-changed-b9948bb7-a208-446d-a45e-2a74affe299b. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1071.651445] env[65385]: DEBUG oslo_concurrency.lockutils [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Acquiring lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.651590] env[65385]: DEBUG oslo_concurrency.lockutils [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Acquired lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.651772] env[65385]: DEBUG nova.network.neutron [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Refreshing network info cache for port b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1071.920688] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454347, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07289} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.921087] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.921768] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5bbafe-a08d-4e6f-b584-c501f8f268dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.945297] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 04ebe83c-cab5-45e1-9f5f-018481f63912/04ebe83c-cab5-45e1-9f5f-018481f63912.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.946065] env[65385]: DEBUG oslo_concurrency.lockutils [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.946323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.946547] env[65385]: INFO nova.compute.manager [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Attaching volume 1b184555-e286-4867-9a99-c29012c47f92 to /dev/sdb [ 1071.948225] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-023f52ba-4946-4fa5-92d8-cc1cbdcb6774 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.973486] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1071.973486] env[65385]: value = "task-4454352" [ 1071.973486] env[65385]: _type = "Task" [ 1071.973486] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.989034] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454352, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.003508] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454349, 'name': Rename_Task, 'duration_secs': 0.183989} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.004269] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1072.004555] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-426e68f4-07fe-4d9b-ad49-2c9bb5d0acc1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.009185] env[65385]: DEBUG oslo_vmware.api [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223586} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.009785] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.010013] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1072.010203] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1072.010375] env[65385]: INFO nova.compute.manager [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Took 1.26 seconds to destroy the instance on the hypervisor. [ 1072.010640] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1072.010853] env[65385]: DEBUG nova.compute.manager [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1072.010952] env[65385]: DEBUG nova.network.neutron [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1072.011234] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.011758] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.012048] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.024175] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4dbacf-8696-448b-813b-5f10487e6b92 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.027118] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1072.027118] env[65385]: value = "task-4454353" [ 1072.027118] env[65385]: _type = "Task" [ 1072.027118] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.037033] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e7cea9-bea4-40bc-9b13-c42fd6f38522 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.042134] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454353, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.055263] env[65385]: DEBUG nova.virt.block_device [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updating existing volume attachment record: f842c97e-6929-4466-8b21-3e6bf1aef263 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1072.060392] env[65385]: DEBUG nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1072.078285] env[65385]: DEBUG oslo_vmware.api [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454351, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291166} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.083183] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.083183] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1072.083183] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1072.083749] env[65385]: INFO nova.compute.manager [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1072.083920] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1072.087653] env[65385]: DEBUG nova.compute.manager [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1072.087653] env[65385]: DEBUG nova.network.neutron [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1072.087653] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.087653] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.087653] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.097247] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.139904] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.156900] env[65385]: WARNING neutronclient.v2_0.client [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.157705] env[65385]: WARNING openstack [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.157977] env[65385]: WARNING openstack [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.185802] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.261302] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd03b8d-0879-4a48-aeea-b19c262c27ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.269838] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d290b4be-4a97-4b4d-b8be-d276ab5f59b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.303608] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edda3d0a-ef39-474e-b085-6b01cdb6435d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.313027] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8d662f-1d10-4b85-9042-4b70a353adad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.331444] env[65385]: DEBUG nova.compute.provider_tree [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.489073] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454352, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.545796] env[65385]: DEBUG oslo_vmware.api [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454353, 'name': PowerOnVM_Task, 'duration_secs': 0.50705} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.546018] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1072.546238] env[65385]: INFO nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Took 8.93 seconds to spawn the instance on the hypervisor. [ 1072.546443] env[65385]: DEBUG nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1072.547503] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85e02ff-99a0-4b66-bdd0-d9c633a705e5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.568790] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.569517] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.596340] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.662733] env[65385]: WARNING openstack [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.663502] env[65385]: WARNING openstack [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.759217] env[65385]: WARNING neutronclient.v2_0.client [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.759952] env[65385]: WARNING openstack [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.760425] env[65385]: WARNING openstack [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.790197] env[65385]: DEBUG nova.compute.manager [req-69875241-5730-43e0-8bd5-7dbf3831b228 req-9f37fb40-6956-42e1-964d-11fc6077e257 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Received event network-vif-deleted-f7b3d6ae-eb2d-43c1-8c44-ca85849be867 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1072.790499] env[65385]: INFO nova.compute.manager [req-69875241-5730-43e0-8bd5-7dbf3831b228 req-9f37fb40-6956-42e1-964d-11fc6077e257 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Neutron deleted interface f7b3d6ae-eb2d-43c1-8c44-ca85849be867; detaching it from the instance and deleting it from the info cache [ 1072.790738] env[65385]: DEBUG nova.network.neutron [req-69875241-5730-43e0-8bd5-7dbf3831b228 req-9f37fb40-6956-42e1-964d-11fc6077e257 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1072.835619] env[65385]: DEBUG nova.scheduler.client.report [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1072.861243] env[65385]: DEBUG nova.network.neutron [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Updated VIF entry in instance network info cache for port b9948bb7-a208-446d-a45e-2a74affe299b. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1072.861383] env[65385]: DEBUG nova.network.neutron [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Updating instance_info_cache with network_info: [{"id": "b9948bb7-a208-446d-a45e-2a74affe299b", "address": "fa:16:3e:87:22:22", "network": {"id": "006c9946-474b-48b9-8af4-c5ae5034f3d6", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2098203597-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87efee8d30ee4a69bdd8dc784e809bfe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9948bb7-a2", "ovs_interfaceid": "b9948bb7-a208-446d-a45e-2a74affe299b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1072.904614] env[65385]: DEBUG nova.network.neutron [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1072.986514] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454352, 'name': ReconfigVM_Task, 'duration_secs': 0.676488} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.986724] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 04ebe83c-cab5-45e1-9f5f-018481f63912/04ebe83c-cab5-45e1-9f5f-018481f63912.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.987434] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6011448-07f9-4fee-9d7f-3e94b4311f4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.998168] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1072.998168] env[65385]: value = "task-4454355" [ 1072.998168] env[65385]: _type = "Task" [ 1072.998168] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.013300] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454355, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.020778] env[65385]: DEBUG nova.compute.manager [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1073.075134] env[65385]: DEBUG nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1073.082096] env[65385]: INFO nova.compute.manager [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Took 14.87 seconds to build instance. [ 1073.176021] env[65385]: DEBUG nova.network.neutron [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1073.294433] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8d6b7d4-6d92-4378-829a-6e23a7728f85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.305653] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95ccc79-f987-42f1-bf93-fcaa2be13c56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.339047] env[65385]: DEBUG nova.compute.manager [req-69875241-5730-43e0-8bd5-7dbf3831b228 req-9f37fb40-6956-42e1-964d-11fc6077e257 service nova] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Detach interface failed, port_id=f7b3d6ae-eb2d-43c1-8c44-ca85849be867, reason: Instance 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1073.340116] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1073.340680] env[65385]: DEBUG nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1073.343507] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.158s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1073.345109] env[65385]: INFO nova.compute.claims [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1073.364780] env[65385]: DEBUG oslo_concurrency.lockutils [req-be284dc2-faa2-43ae-bb80-8919df506bf6 req-1eb6c283-7e20-439f-ac1e-ae09623c9240 service nova] Releasing lock "refresh_cache-b1abf6ab-a3d5-4078-9568-68db0a7f03cc" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1073.407324] env[65385]: INFO nova.compute.manager [-] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Took 1.32 seconds to deallocate network for instance. [ 1073.509054] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454355, 'name': Rename_Task, 'duration_secs': 0.264524} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.509379] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.509627] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6993497b-5702-498b-9ef5-080de80df9d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.516689] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1073.516689] env[65385]: value = "task-4454356" [ 1073.516689] env[65385]: _type = "Task" [ 1073.516689] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.529572] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454356, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.558588] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.586390] env[65385]: DEBUG oslo_concurrency.lockutils [None req-da0536b3-1052-4eb2-a831-30a14078f15c tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.394s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1073.602379] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.680654] env[65385]: INFO nova.compute.manager [-] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Took 1.67 seconds to deallocate network for instance. [ 1073.687708] env[65385]: DEBUG nova.compute.manager [req-da75adb7-c4a0-4b5c-b8e5-2fe618237409 req-892ab593-238c-4b92-bf7e-e319adb60b07 service nova] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Received event network-vif-deleted-0fc23e28-1265-46b7-aa43-fd72e2269330 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1073.850342] env[65385]: DEBUG nova.compute.utils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1073.854384] env[65385]: DEBUG nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1073.854384] env[65385]: DEBUG nova.network.neutron [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1073.854557] env[65385]: WARNING neutronclient.v2_0.client [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.854928] env[65385]: WARNING neutronclient.v2_0.client [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.855879] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.856370] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.907642] env[65385]: DEBUG nova.policy [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1073.915629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.030198] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454356, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.063514] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce183c71-748a-4961-b735-900cc8a8ee89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.071817] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc74b81-5114-4460-b770-a29d176c7f28 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.104148] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cccd0a3-dc21-4afa-8070-36c599f72998 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.112920] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d276f026-e44e-4414-ac1b-f8aa1d8786c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.127304] env[65385]: DEBUG nova.compute.provider_tree [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.189982] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.224151] env[65385]: DEBUG nova.network.neutron [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Successfully created port: 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1074.355501] env[65385]: DEBUG nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1074.426048] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.426556] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.426726] env[65385]: DEBUG nova.compute.manager [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1074.427880] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6537298d-04ee-46cc-9f6f-2103a12c11c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.437275] env[65385]: DEBUG nova.compute.manager [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1074.438012] env[65385]: DEBUG nova.objects.instance [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'flavor' on Instance uuid 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.528674] env[65385]: DEBUG oslo_vmware.api [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454356, 'name': PowerOnVM_Task, 'duration_secs': 0.997419} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.528949] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.529169] env[65385]: INFO nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Took 8.59 seconds to spawn the instance on the hypervisor. [ 1074.529340] env[65385]: DEBUG nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1074.530207] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8585d4-10a4-4347-8292-411453bdb43a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.633357] env[65385]: DEBUG nova.scheduler.client.report [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1075.050975] env[65385]: INFO nova.compute.manager [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Took 15.17 seconds to build instance. [ 1075.140173] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.795s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.140173] env[65385]: DEBUG nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1075.142687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.547s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.144231] env[65385]: INFO nova.compute.claims [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1075.370838] env[65385]: DEBUG nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1075.400563] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1075.400801] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1075.400953] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1075.401155] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1075.401305] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1075.401440] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1075.401768] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.401933] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1075.402129] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1075.402282] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1075.402450] env[65385]: DEBUG nova.virt.hardware [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1075.403352] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf34991-7a5b-4171-9716-de158db4106f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.412304] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2adb9d1-d270-40bb-a83a-c9855fb22f4c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.445539] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.445845] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b90f72f-84e1-4019-8986-310f7204de75 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.453734] env[65385]: DEBUG oslo_vmware.api [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1075.453734] env[65385]: value = "task-4454358" [ 1075.453734] env[65385]: _type = "Task" [ 1075.453734] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.467413] env[65385]: DEBUG oslo_vmware.api [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454358, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.553563] env[65385]: DEBUG oslo_concurrency.lockutils [None req-55d99ff2-8392-4408-a58d-594d585ec0db tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.681s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.648705] env[65385]: DEBUG nova.compute.utils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1075.653418] env[65385]: DEBUG nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1075.653625] env[65385]: DEBUG nova.network.neutron [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1075.654016] env[65385]: WARNING neutronclient.v2_0.client [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.655051] env[65385]: WARNING neutronclient.v2_0.client [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.655966] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.656452] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.688614] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "04ebe83c-cab5-45e1-9f5f-018481f63912" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.689781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.689781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "04ebe83c-cab5-45e1-9f5f-018481f63912-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.689781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.689781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.694538] env[65385]: INFO nova.compute.manager [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Terminating instance [ 1075.721122] env[65385]: DEBUG nova.policy [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2a287457352470794887b229ea90cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60de0d1c162342209795a180391661f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1075.729260] env[65385]: DEBUG nova.compute.manager [req-0fad4a67-5e8d-4acb-b6ed-e041100c7dc7 req-1929a304-04a4-4180-93c1-62f26eb5ff58 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-vif-plugged-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1075.729474] env[65385]: DEBUG oslo_concurrency.lockutils [req-0fad4a67-5e8d-4acb-b6ed-e041100c7dc7 req-1929a304-04a4-4180-93c1-62f26eb5ff58 service nova] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.729671] env[65385]: DEBUG oslo_concurrency.lockutils [req-0fad4a67-5e8d-4acb-b6ed-e041100c7dc7 req-1929a304-04a4-4180-93c1-62f26eb5ff58 service nova] Lock "900639b6-9b98-436f-aaad-ea8391f67393-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.729829] env[65385]: DEBUG oslo_concurrency.lockutils [req-0fad4a67-5e8d-4acb-b6ed-e041100c7dc7 req-1929a304-04a4-4180-93c1-62f26eb5ff58 service nova] Lock "900639b6-9b98-436f-aaad-ea8391f67393-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.730021] env[65385]: DEBUG nova.compute.manager [req-0fad4a67-5e8d-4acb-b6ed-e041100c7dc7 req-1929a304-04a4-4180-93c1-62f26eb5ff58 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] No waiting events found dispatching network-vif-plugged-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1075.730263] env[65385]: WARNING nova.compute.manager [req-0fad4a67-5e8d-4acb-b6ed-e041100c7dc7 req-1929a304-04a4-4180-93c1-62f26eb5ff58 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received unexpected event network-vif-plugged-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 for instance with vm_state building and task_state spawning. [ 1075.795517] env[65385]: DEBUG nova.network.neutron [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Successfully updated port: 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1075.884526] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba153c5-7314-4cd3-9b1a-3f70e5774dbc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.893146] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a681fd7-a680-4992-9e64-42ba2c8daaf2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.925146] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6182bdf0-2823-4750-94d0-e3f9f0a135dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.934679] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60438666-64b1-4f99-afe3-886829543df3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.950192] env[65385]: DEBUG nova.compute.provider_tree [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.963932] env[65385]: DEBUG oslo_vmware.api [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454358, 'name': PowerOffVM_Task, 'duration_secs': 0.193731} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.964301] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.964532] env[65385]: DEBUG nova.compute.manager [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1075.965437] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c68e41b-af51-4919-9250-13a9e102c86b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.056153] env[65385]: DEBUG nova.network.neutron [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Successfully created port: 92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1076.152261] env[65385]: DEBUG nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1076.202254] env[65385]: DEBUG nova.compute.manager [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1076.202533] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1076.203461] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d11251-322c-4f0d-8bd3-0463ab8a69fd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.212546] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1076.212808] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-185973ca-78b8-438e-b18f-ce5bb1a410e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.222164] env[65385]: DEBUG oslo_vmware.api [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1076.222164] env[65385]: value = "task-4454359" [ 1076.222164] env[65385]: _type = "Task" [ 1076.222164] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.232735] env[65385]: DEBUG oslo_vmware.api [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.303217] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.303439] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.303589] env[65385]: DEBUG nova.network.neutron [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1076.453750] env[65385]: DEBUG nova.scheduler.client.report [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1076.483172] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3b7852bf-257d-46d0-9297-90dc95780d92 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.057s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.618020] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1076.618020] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871167', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'name': 'volume-1b184555-e286-4867-9a99-c29012c47f92', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5470afe7-6dd0-4778-9b83-2956ede04b7f', 'attached_at': '', 'detached_at': '', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'serial': '1b184555-e286-4867-9a99-c29012c47f92'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1076.619467] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce27a008-4cfe-4217-90c4-bf1cff87fd02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.639081] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f1619e-bff3-4f33-827f-3841a1ec6af9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.670917] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-1b184555-e286-4867-9a99-c29012c47f92/volume-1b184555-e286-4867-9a99-c29012c47f92.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.671544] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c797c5d9-11fc-4700-b268-838c4264bf0c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.696767] env[65385]: DEBUG oslo_vmware.api [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1076.696767] env[65385]: value = "task-4454360" [ 1076.696767] env[65385]: _type = "Task" [ 1076.696767] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.706975] env[65385]: DEBUG oslo_vmware.api [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454360, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.732979] env[65385]: DEBUG oslo_vmware.api [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454359, 'name': PowerOffVM_Task, 'duration_secs': 0.30936} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.733360] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1076.733583] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1076.733920] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12baa30c-f580-491d-9a2f-4e21d3caf5ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.808512] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.808984] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.817994] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1076.818239] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1076.818420] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleting the datastore file [datastore2] 04ebe83c-cab5-45e1-9f5f-018481f63912 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1076.818924] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98b366bd-9e35-429e-b5cf-d2398f868ec3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.828668] env[65385]: DEBUG oslo_vmware.api [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1076.828668] env[65385]: value = "task-4454362" [ 1076.828668] env[65385]: _type = "Task" [ 1076.828668] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.837768] env[65385]: DEBUG oslo_vmware.api [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454362, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.873168] env[65385]: DEBUG nova.network.neutron [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1076.920079] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.920624] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.960695] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.818s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.961168] env[65385]: DEBUG nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1076.963960] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.405s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1077.044177] env[65385]: WARNING neutronclient.v2_0.client [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.045275] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.046278] env[65385]: WARNING openstack [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.171153] env[65385]: DEBUG nova.network.neutron [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.172879] env[65385]: DEBUG nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1077.197162] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1077.197668] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1077.197999] env[65385]: DEBUG nova.virt.hardware [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1077.199128] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b3adeb-9b9c-4557-8165-98552fcff22f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.219053] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a0a683-197d-4746-b647-7e88ea4b408c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.223880] env[65385]: DEBUG oslo_vmware.api [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454360, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.339768] env[65385]: DEBUG oslo_vmware.api [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454362, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209835} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.340089] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1077.340283] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1077.340456] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1077.340627] env[65385]: INFO nova.compute.manager [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1077.340867] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1077.341147] env[65385]: DEBUG nova.compute.manager [-] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1077.341369] env[65385]: DEBUG nova.network.neutron [-] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1077.341478] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.342017] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.342283] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.354682] env[65385]: INFO nova.compute.manager [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Rebuilding instance [ 1077.380284] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.406975] env[65385]: DEBUG nova.compute.manager [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1077.407772] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eccae2b9-5e6c-4de4-8ab7-6d28b5328760 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.467647] env[65385]: DEBUG nova.compute.utils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1077.472883] env[65385]: INFO nova.compute.claims [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.475343] env[65385]: DEBUG nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1077.475586] env[65385]: DEBUG nova.network.neutron [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1077.475968] env[65385]: WARNING neutronclient.v2_0.client [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.476338] env[65385]: WARNING neutronclient.v2_0.client [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.476910] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.477322] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.606408] env[65385]: DEBUG nova.network.neutron [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Successfully updated port: 92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1077.661161] env[65385]: DEBUG nova.policy [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c1a9d4194964403a8d11abfad4c65a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94cc5bc221c0455ea760f5022db6bdc4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1077.665585] env[65385]: DEBUG nova.compute.manager [req-1a716773-1911-4b92-8052-8aab1ecde98e req-c434de1d-e79e-40b0-9155-be8ee5fecc1a service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Received event network-vif-deleted-32b4dfff-31d2-41e4-9509-668e6d7febf5 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1077.665853] env[65385]: INFO nova.compute.manager [req-1a716773-1911-4b92-8052-8aab1ecde98e req-c434de1d-e79e-40b0-9155-be8ee5fecc1a service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Neutron deleted interface 32b4dfff-31d2-41e4-9509-668e6d7febf5; detaching it from the instance and deleting it from the info cache [ 1077.666042] env[65385]: DEBUG nova.network.neutron [req-1a716773-1911-4b92-8052-8aab1ecde98e req-c434de1d-e79e-40b0-9155-be8ee5fecc1a service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.673018] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.673370] env[65385]: DEBUG nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Instance network_info: |[{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1077.674055] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:b2:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1077.682676] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1077.683698] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1077.684486] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97cb663c-8239-4ccd-bce9-7b47e59fc29a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.709574] env[65385]: DEBUG oslo_vmware.api [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454360, 'name': ReconfigVM_Task, 'duration_secs': 0.531744} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.710963] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-1b184555-e286-4867-9a99-c29012c47f92/volume-1b184555-e286-4867-9a99-c29012c47f92.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.717017] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1077.717017] env[65385]: value = "task-4454363" [ 1077.717017] env[65385]: _type = "Task" [ 1077.717017] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.717017] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e0b7baf-ce6e-442e-a8c2-a346068e9acc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.740025] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454363, 'name': CreateVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.740025] env[65385]: DEBUG oslo_vmware.api [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1077.740025] env[65385]: value = "task-4454364" [ 1077.740025] env[65385]: _type = "Task" [ 1077.740025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.750974] env[65385]: DEBUG oslo_vmware.api [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454364, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.761884] env[65385]: DEBUG nova.compute.manager [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1077.762321] env[65385]: DEBUG nova.compute.manager [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing instance network info cache due to event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1077.762446] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.762498] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.762638] env[65385]: DEBUG nova.network.neutron [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1077.958804] env[65385]: DEBUG nova.network.neutron [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Successfully created port: ed7c17c7-54b1-4cee-9cb2-139730bbf715 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1077.976051] env[65385]: DEBUG nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1077.987503] env[65385]: INFO nova.compute.resource_tracker [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating resource usage from migration d3c745d8-6882-4477-bdda-3c679cb5ee1b [ 1078.091115] env[65385]: DEBUG nova.network.neutron [-] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.112943] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.112943] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.112943] env[65385]: DEBUG nova.network.neutron [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1078.170335] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01b1958a-2fff-4a6f-a42e-21be79ea7d58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.183717] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ffc3d3-93b6-4351-9606-f4432bfe5363 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.220027] env[65385]: DEBUG nova.compute.manager [req-1a716773-1911-4b92-8052-8aab1ecde98e req-c434de1d-e79e-40b0-9155-be8ee5fecc1a service nova] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Detach interface failed, port_id=32b4dfff-31d2-41e4-9509-668e6d7febf5, reason: Instance 04ebe83c-cab5-45e1-9f5f-018481f63912 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1078.236465] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454363, 'name': CreateVM_Task, 'duration_secs': 0.369533} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.236650] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1078.239595] env[65385]: WARNING neutronclient.v2_0.client [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.240358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.240358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.240508] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1078.240978] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8d622f8-fe4c-4e29-928b-0ab74e8512d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.251517] env[65385]: DEBUG oslo_vmware.api [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454364, 'name': ReconfigVM_Task, 'duration_secs': 0.162403} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.253872] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871167', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'name': 'volume-1b184555-e286-4867-9a99-c29012c47f92', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5470afe7-6dd0-4778-9b83-2956ede04b7f', 'attached_at': '', 'detached_at': '', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'serial': '1b184555-e286-4867-9a99-c29012c47f92'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1078.256101] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1078.256101] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520019ca-c8b7-566c-01a8-909b05d8e766" [ 1078.256101] env[65385]: _type = "Task" [ 1078.256101] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.256433] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabc62c5-77e0-4aa5-87af-9a7da65d6523 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.267783] env[65385]: WARNING neutronclient.v2_0.client [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.268410] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.268760] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.281453] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02f91f9-fee6-4d57-ab65-5d5194eb960b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.285081] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520019ca-c8b7-566c-01a8-909b05d8e766, 'name': SearchDatastore_Task, 'duration_secs': 0.01473} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.285715] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.286182] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.286700] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.286700] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.286700] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.287280] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02121890-f825-4f95-bbab-038804a28ae1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.317591] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6f0827-b539-4735-9e8f-3c46bebcf4f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.320377] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.320562] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1078.321436] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b28149f2-c5cd-4b67-a9d0-c9c363781186 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.334262] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5991bc-92bd-4f57-8b5f-27854b646d09 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.338815] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1078.338815] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dc6302-e980-1ebe-af59-92908f4ab3f0" [ 1078.338815] env[65385]: _type = "Task" [ 1078.338815] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.351992] env[65385]: DEBUG nova.compute.provider_tree [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.360468] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dc6302-e980-1ebe-af59-92908f4ab3f0, 'name': SearchDatastore_Task, 'duration_secs': 0.010427} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.361722] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43690f20-298e-42bb-8d11-4c1e4793e424 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.369119] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1078.369119] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523f956c-b4f3-b354-8d2b-8d8948e15fb1" [ 1078.369119] env[65385]: _type = "Task" [ 1078.369119] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.381791] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523f956c-b4f3-b354-8d2b-8d8948e15fb1, 'name': SearchDatastore_Task, 'duration_secs': 0.010461} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.382080] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.382349] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 900639b6-9b98-436f-aaad-ea8391f67393/900639b6-9b98-436f-aaad-ea8391f67393.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1078.382619] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fe084c8-3d68-4a2b-a9d7-8f76c186c5f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.391273] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1078.391273] env[65385]: value = "task-4454365" [ 1078.391273] env[65385]: _type = "Task" [ 1078.391273] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.400512] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.424269] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.424645] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-101bea15-7ae2-41d3-a81d-3d4a9a6e8bb7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.427911] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.428371] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.442823] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1078.442823] env[65385]: value = "task-4454366" [ 1078.442823] env[65385]: _type = "Task" [ 1078.442823] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.454470] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1078.454795] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1078.455749] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fd69f2-59fc-4efa-a9ab-3e9b3b8631b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.469528] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1078.469878] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90e8f822-239d-4438-8839-52e0f53d99e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.504273] env[65385]: WARNING neutronclient.v2_0.client [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.505143] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.505527] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.544302] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1078.544668] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1078.545011] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleting the datastore file [datastore1] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.545307] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7605e78-094c-4b72-be07-79d9fb812fac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.554314] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1078.554314] env[65385]: value = "task-4454368" [ 1078.554314] env[65385]: _type = "Task" [ 1078.554314] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.564951] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.598667] env[65385]: INFO nova.compute.manager [-] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Took 1.26 seconds to deallocate network for instance. [ 1078.615508] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.615966] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.625969] env[65385]: DEBUG nova.network.neutron [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updated VIF entry in instance network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1078.626150] env[65385]: DEBUG nova.network.neutron [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.667070] env[65385]: DEBUG nova.network.neutron [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1078.689357] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.689882] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.855827] env[65385]: DEBUG nova.scheduler.client.report [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.860102] env[65385]: WARNING neutronclient.v2_0.client [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.860831] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.861521] env[65385]: WARNING openstack [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.903334] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474233} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.903334] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 900639b6-9b98-436f-aaad-ea8391f67393/900639b6-9b98-436f-aaad-ea8391f67393.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1078.903334] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.903502] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c5ad54c-17ac-4f45-a361-ca12f7a992ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.912940] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1078.912940] env[65385]: value = "task-4454369" [ 1078.912940] env[65385]: _type = "Task" [ 1078.912940] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.924211] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454369, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.986795] env[65385]: DEBUG nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1078.991096] env[65385]: DEBUG nova.network.neutron [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1079.027216] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1079.027485] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1079.027636] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1079.027811] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1079.027952] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1079.028155] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1079.028429] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1079.028590] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1079.028769] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1079.028931] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1079.029114] env[65385]: DEBUG nova.virt.hardware [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1079.030128] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27df6f73-430d-4dbf-98a3-8225a102a206 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.039153] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49dd19a1-e8e4-424b-9467-db0ffe267444 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.065894] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.382643} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.066190] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.066368] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1079.066539] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1079.106068] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.129552] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.129607] env[65385]: DEBUG nova.compute.manager [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Received event network-vif-plugged-92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1079.129763] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Acquiring lock "6c826914-4c6b-456b-a403-2373d30c803f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.129970] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Lock "6c826914-4c6b-456b-a403-2373d30c803f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.130144] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Lock "6c826914-4c6b-456b-a403-2373d30c803f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.130307] env[65385]: DEBUG nova.compute.manager [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] No waiting events found dispatching network-vif-plugged-92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1079.130469] env[65385]: WARNING nova.compute.manager [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Received unexpected event network-vif-plugged-92c45f41-92b4-40c3-8516-30189d29e33d for instance with vm_state building and task_state spawning. [ 1079.130623] env[65385]: DEBUG nova.compute.manager [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Received event network-changed-92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1079.130772] env[65385]: DEBUG nova.compute.manager [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Refreshing instance network info cache due to event network-changed-92c45f41-92b4-40c3-8516-30189d29e33d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1079.130929] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Acquiring lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.308569] env[65385]: DEBUG nova.objects.instance [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'flavor' on Instance uuid 5470afe7-6dd0-4778-9b83-2956ede04b7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.372228] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.408s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.372228] env[65385]: INFO nova.compute.manager [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Migrating [ 1079.378845] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.777s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.380785] env[65385]: INFO nova.compute.claims [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1079.424914] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454369, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166739} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.426244] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.427335] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6bae809-e8ad-454f-a6aa-9d6a7b30e377 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.452414] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 900639b6-9b98-436f-aaad-ea8391f67393/900639b6-9b98-436f-aaad-ea8391f67393.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.453020] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12a31287-c5f7-4fee-8c4f-64336c2691c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.475845] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1079.475845] env[65385]: value = "task-4454370" [ 1079.475845] env[65385]: _type = "Task" [ 1079.475845] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.485057] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.493857] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.494343] env[65385]: DEBUG nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Instance network_info: |[{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1079.494952] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Acquired lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.494952] env[65385]: DEBUG nova.network.neutron [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Refreshing network info cache for port 92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1079.496345] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:19:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92c45f41-92b4-40c3-8516-30189d29e33d', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1079.504372] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1079.508254] env[65385]: DEBUG nova.network.neutron [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Successfully updated port: ed7c17c7-54b1-4cee-9cb2-139730bbf715 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1079.508254] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1079.508254] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13e250ab-b6b9-46f2-9184-0da6bc245291 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.531147] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1079.531147] env[65385]: value = "task-4454371" [ 1079.531147] env[65385]: _type = "Task" [ 1079.531147] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.541114] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454371, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.805800] env[65385]: DEBUG nova.compute.manager [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Received event network-vif-plugged-ed7c17c7-54b1-4cee-9cb2-139730bbf715 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1079.805934] env[65385]: DEBUG oslo_concurrency.lockutils [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Acquiring lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.806271] env[65385]: DEBUG oslo_concurrency.lockutils [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.806498] env[65385]: DEBUG oslo_concurrency.lockutils [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.806692] env[65385]: DEBUG nova.compute.manager [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] No waiting events found dispatching network-vif-plugged-ed7c17c7-54b1-4cee-9cb2-139730bbf715 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1079.806887] env[65385]: WARNING nova.compute.manager [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Received unexpected event network-vif-plugged-ed7c17c7-54b1-4cee-9cb2-139730bbf715 for instance with vm_state building and task_state spawning. [ 1079.807062] env[65385]: DEBUG nova.compute.manager [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Received event network-changed-ed7c17c7-54b1-4cee-9cb2-139730bbf715 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1079.807171] env[65385]: DEBUG nova.compute.manager [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Refreshing instance network info cache due to event network-changed-ed7c17c7-54b1-4cee-9cb2-139730bbf715. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1079.807421] env[65385]: DEBUG oslo_concurrency.lockutils [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Acquiring lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.807505] env[65385]: DEBUG oslo_concurrency.lockutils [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Acquired lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.807677] env[65385]: DEBUG nova.network.neutron [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Refreshing network info cache for port ed7c17c7-54b1-4cee-9cb2-139730bbf715 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1079.812934] env[65385]: DEBUG oslo_concurrency.lockutils [None req-01ad9948-4a6a-4b29-9099-f94f2f5006c7 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.867s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.893108] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.893330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.893538] env[65385]: DEBUG nova.network.neutron [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1079.985935] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454370, 'name': ReconfigVM_Task, 'duration_secs': 0.424154} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.986254] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 900639b6-9b98-436f-aaad-ea8391f67393/900639b6-9b98-436f-aaad-ea8391f67393.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.986899] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2fb89e5a-ebfc-4717-a3c3-f1e6dae81e8d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.996503] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1079.996503] env[65385]: value = "task-4454372" [ 1079.996503] env[65385]: _type = "Task" [ 1079.996503] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.008927] env[65385]: WARNING neutronclient.v2_0.client [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.009810] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.010031] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.018775] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.019166] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454372, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.042799] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454371, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.107222] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1080.107406] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1080.107608] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1080.107835] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1080.108025] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1080.108192] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1080.108460] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.108642] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1080.108836] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1080.109044] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1080.109264] env[65385]: DEBUG nova.virt.hardware [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1080.110219] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee636ea-2de4-43bd-a638-de6ee45e487f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.127432] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8176aed-f47f-42e9-a278-97c5715783f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.146597] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:73:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd33bca4a-a999-48ad-8138-b0f8d5b6ee23', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1080.155587] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1080.157265] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.157647] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.165075] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1080.165624] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bdcb4e5-3fab-45a7-9d47-9f1de61f2ccd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.188492] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1080.188492] env[65385]: value = "task-4454373" [ 1080.188492] env[65385]: _type = "Task" [ 1080.188492] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.202687] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454373, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.228196] env[65385]: WARNING neutronclient.v2_0.client [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.229036] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.229314] env[65385]: WARNING openstack [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.311891] env[65385]: WARNING neutronclient.v2_0.client [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.312362] env[65385]: WARNING openstack [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.312737] env[65385]: WARNING openstack [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.321671] env[65385]: DEBUG nova.network.neutron [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updated VIF entry in instance network info cache for port 92c45f41-92b4-40c3-8516-30189d29e33d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1080.322174] env[65385]: DEBUG nova.network.neutron [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1080.353481] env[65385]: DEBUG nova.network.neutron [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1080.401059] env[65385]: WARNING neutronclient.v2_0.client [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.401059] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.401389] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.480488] env[65385]: DEBUG nova.network.neutron [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1080.515136] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454372, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.547420] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454371, 'name': CreateVM_Task, 'duration_secs': 0.931442} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.548898] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.549266] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.556169] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1080.556962] env[65385]: WARNING neutronclient.v2_0.client [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.557438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.557601] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1080.557923] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1080.560733] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78c42f77-c0c2-429a-b119-4b8283df5834 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.567778] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1080.567778] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a5c808-ef3c-bd82-9ca9-d61e73b41d63" [ 1080.567778] env[65385]: _type = "Task" [ 1080.567778] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.580861] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a5c808-ef3c-bd82-9ca9-d61e73b41d63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.628675] env[65385]: DEBUG oslo_concurrency.lockutils [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1080.628922] env[65385]: DEBUG oslo_concurrency.lockutils [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1080.631419] env[65385]: WARNING neutronclient.v2_0.client [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.631901] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.632283] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.641379] env[65385]: INFO nova.compute.manager [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Detaching volume 1b184555-e286-4867-9a99-c29012c47f92 [ 1080.653310] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37206ef-b03b-4b7f-adf6-fe238a853d34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.662189] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b668a62a-4c20-4b4e-8aba-452699bd8260 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.700582] env[65385]: INFO nova.virt.block_device [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Attempting to driver detach volume 1b184555-e286-4867-9a99-c29012c47f92 from mountpoint /dev/sdb [ 1080.700835] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1080.701035] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871167', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'name': 'volume-1b184555-e286-4867-9a99-c29012c47f92', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5470afe7-6dd0-4778-9b83-2956ede04b7f', 'attached_at': '', 'detached_at': '', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'serial': '1b184555-e286-4867-9a99-c29012c47f92'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1080.704563] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92022b50-2726-4dad-9f56-0a3b1675cf47 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.707614] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7a51a2-1c91-4a12-94ff-c76020ad8469 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.734135] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c358889-5ae0-4ea6-af44-35110f01b250 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.738339] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454373, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.739498] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c46012-5eb2-4763-b831-436965670444 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.754855] env[65385]: DEBUG nova.compute.provider_tree [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.758234] env[65385]: DEBUG nova.network.neutron [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1080.761758] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fdee88-cf97-4d44-9301-c714f547b370 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.784190] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ba6c22-6711-4afe-b760-61fe3c9d79e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.800676] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The volume has not been displaced from its original location: [datastore2] volume-1b184555-e286-4867-9a99-c29012c47f92/volume-1b184555-e286-4867-9a99-c29012c47f92.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1080.805946] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1080.806259] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d16c39bd-d318-4271-b1cc-5d75fd152dd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.825667] env[65385]: DEBUG oslo_concurrency.lockutils [req-06621f69-46ad-46b5-8bae-de80bbaf3251 req-160c5590-2977-4f2b-8ca8-da7500cd6956 service nova] Releasing lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.826276] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1080.826276] env[65385]: value = "task-4454374" [ 1080.826276] env[65385]: _type = "Task" [ 1080.826276] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.836329] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454374, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.983037] env[65385]: DEBUG oslo_concurrency.lockutils [req-fb44a325-cb99-4528-9f7b-2ba83c31c672 req-79e55e17-7e26-4837-af01-9a36c581b47f service nova] Releasing lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.983037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1080.983363] env[65385]: DEBUG nova.network.neutron [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1081.013199] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454372, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.079499] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a5c808-ef3c-bd82-9ca9-d61e73b41d63, 'name': SearchDatastore_Task, 'duration_secs': 0.011934} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.079866] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1081.080209] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1081.080495] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.080664] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.080848] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.081161] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dab6d9a4-f76c-4aaf-a932-8e46f908d041 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.090998] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.091221] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1081.091970] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62acdd3f-72a5-4612-b47d-ea7d867de6a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.097798] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1081.097798] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d125f-d80f-4242-74ab-6bcb97fd6070" [ 1081.097798] env[65385]: _type = "Task" [ 1081.097798] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.106090] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d125f-d80f-4242-74ab-6bcb97fd6070, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.206014] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454373, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.262798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1081.265226] env[65385]: DEBUG nova.scheduler.client.report [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1081.338403] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454374, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.486683] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.486683] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.515590] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454372, 'name': Rename_Task, 'duration_secs': 1.203703} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.515717] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1081.515996] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f1d9f2b-3956-4800-8d9d-cddf246bd2d5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.523485] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1081.523485] env[65385]: value = "task-4454375" [ 1081.523485] env[65385]: _type = "Task" [ 1081.523485] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.524505] env[65385]: DEBUG nova.network.neutron [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1081.535788] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454375, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.554277] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.554822] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.608358] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525d125f-d80f-4242-74ab-6bcb97fd6070, 'name': SearchDatastore_Task, 'duration_secs': 0.012195} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.609191] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-803fd3b8-853c-410e-bb00-626522b6f31c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.615018] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1081.615018] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa5826-b46c-1cc9-0f18-e9e87440b895" [ 1081.615018] env[65385]: _type = "Task" [ 1081.615018] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.623904] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa5826-b46c-1cc9-0f18-e9e87440b895, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.634355] env[65385]: WARNING neutronclient.v2_0.client [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.635027] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.635374] env[65385]: WARNING openstack [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.706864] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454373, 'name': CreateVM_Task, 'duration_secs': 1.259973} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.707096] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1081.707615] env[65385]: WARNING neutronclient.v2_0.client [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1081.707982] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.708265] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.708460] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1081.708729] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec3071f9-8b8e-4038-84c5-f6b9152c06df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.714384] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1081.714384] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cadf01-4b24-89a0-3c11-53daadc353ef" [ 1081.714384] env[65385]: _type = "Task" [ 1081.714384] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.723430] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cadf01-4b24-89a0-3c11-53daadc353ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.725893] env[65385]: DEBUG nova.network.neutron [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance_info_cache with network_info: [{"id": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "address": "fa:16:3e:47:9a:0a", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped7c17c7-54", "ovs_interfaceid": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1081.775933] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.776417] env[65385]: DEBUG nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1081.779321] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.864s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.779543] env[65385]: DEBUG nova.objects.instance [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lazy-loading 'resources' on Instance uuid 77d718c8-4959-483a-8717-c00f2be1d0bb {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.837347] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454374, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.037269] env[65385]: DEBUG oslo_vmware.api [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454375, 'name': PowerOnVM_Task, 'duration_secs': 0.474739} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.037542] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1082.037740] env[65385]: INFO nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1082.037916] env[65385]: DEBUG nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1082.038702] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48592090-ecf3-4be0-9fcb-bdc73881c07f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.126509] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa5826-b46c-1cc9-0f18-e9e87440b895, 'name': SearchDatastore_Task, 'duration_secs': 0.018267} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.126757] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.127044] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1082.127339] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be6d1374-1e25-4da3-9e99-9beb74d1dba7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.135881] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1082.135881] env[65385]: value = "task-4454376" [ 1082.135881] env[65385]: _type = "Task" [ 1082.135881] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.145453] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454376, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.226915] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cadf01-4b24-89a0-3c11-53daadc353ef, 'name': SearchDatastore_Task, 'duration_secs': 0.012334} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.227234] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.227482] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.227730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.227872] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.228770] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.228770] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.228889] env[65385]: DEBUG nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Instance network_info: |[{"id": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "address": "fa:16:3e:47:9a:0a", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped7c17c7-54", "ovs_interfaceid": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1082.229109] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4f35921-6be0-45c5-b063-7ffd75348436 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.231290] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:9a:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed7c17c7-54b1-4cee-9cb2-139730bbf715', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.238712] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1082.239382] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1082.239673] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c53f9268-848d-471a-ae1d-fc50b4ecf6cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.257862] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.258118] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1082.259356] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-107bcb61-7f51-4b2a-9b1e-a27e24e20b61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.265229] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.265229] env[65385]: value = "task-4454377" [ 1082.265229] env[65385]: _type = "Task" [ 1082.265229] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.266727] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1082.266727] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c1206c-11b3-e4e7-7699-ece2e1fff814" [ 1082.266727] env[65385]: _type = "Task" [ 1082.266727] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.281909] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454377, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.287184] env[65385]: DEBUG nova.compute.utils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1082.291430] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c1206c-11b3-e4e7-7699-ece2e1fff814, 'name': SearchDatastore_Task, 'duration_secs': 0.009868} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.292034] env[65385]: DEBUG nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1082.292240] env[65385]: DEBUG nova.network.neutron [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1082.292718] env[65385]: WARNING neutronclient.v2_0.client [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.292929] env[65385]: WARNING neutronclient.v2_0.client [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.293575] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.293988] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.303825] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76610613-b336-49d9-bda5-525d883f3bdb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.311125] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1082.311125] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b98f50-86bc-ceae-246c-85902f6e2057" [ 1082.311125] env[65385]: _type = "Task" [ 1082.311125] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.326355] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b98f50-86bc-ceae-246c-85902f6e2057, 'name': SearchDatastore_Task, 'duration_secs': 0.009674} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.327176] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.327501] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1082.327928] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54e36839-9c6f-4d73-9d57-aaa9a4b2bd95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.345908] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454374, 'name': ReconfigVM_Task, 'duration_secs': 1.377611} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.347798] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1082.352965] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1082.352965] env[65385]: value = "task-4454378" [ 1082.352965] env[65385]: _type = "Task" [ 1082.352965] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.356203] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ab74fae-7870-497a-842d-43d452a5a90c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.371058] env[65385]: DEBUG nova.policy [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd174ab43a1654f169c7df2d3cf73f52a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a282a49e5f44dc781471d9c2103ca8f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1082.387098] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1082.387098] env[65385]: value = "task-4454379" [ 1082.387098] env[65385]: _type = "Task" [ 1082.387098] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.387415] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454378, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.397884] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454379, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.558696] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f476acdd-0cd9-431c-b568-9c29e6776a66 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.564094] env[65385]: INFO nova.compute.manager [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Took 11.59 seconds to build instance. [ 1082.570732] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1076b895-4ab4-4ce5-b97e-02efe07e53f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.623274] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880d0045-cc04-476d-bb06-e5a8b0997abc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.634654] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7e627f-467f-4996-8d1b-863f0df19287 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.656456] env[65385]: DEBUG nova.compute.provider_tree [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.661260] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454376, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.772623] env[65385]: DEBUG nova.network.neutron [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Successfully created port: bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1082.782191] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454377, 'name': CreateVM_Task, 'duration_secs': 0.443647} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.785041] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1082.785724] env[65385]: WARNING neutronclient.v2_0.client [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.786550] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.786621] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.787152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1082.788707] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34c55aa0-ab3b-48d4-9f31-dffbfcff9f5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.792194] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131500e5-46ea-4c0f-9ba8-61707adf5af7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.797940] env[65385]: DEBUG nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1082.824736] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance '29a449d0-f62d-4430-96ab-9ac43e5a7ad2' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1082.830673] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1082.830673] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b69f07-3dd4-55f8-fd8e-01ca740bb64f" [ 1082.830673] env[65385]: _type = "Task" [ 1082.830673] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.849521] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b69f07-3dd4-55f8-fd8e-01ca740bb64f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.881100] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454378, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.901639] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454379, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.066298] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fbc4f1de-7330-4a23-b3ae-834e2ce33892 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "900639b6-9b98-436f-aaad-ea8391f67393" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.105s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.152382] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454376, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54849} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.152739] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1083.153680] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.154176] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62c42b31-70de-4a18-8f27-21d8e8ef0677 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.161987] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1083.161987] env[65385]: value = "task-4454380" [ 1083.161987] env[65385]: _type = "Task" [ 1083.161987] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.163024] env[65385]: DEBUG nova.scheduler.client.report [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1083.179079] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.334076] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.334211] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1007634d-6e79-4df4-aa59-eb4440da1a45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.349582] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b69f07-3dd4-55f8-fd8e-01ca740bb64f, 'name': SearchDatastore_Task, 'duration_secs': 0.060537} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.351497] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.351815] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.352127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.353617] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.354048] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.354546] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1083.354546] env[65385]: value = "task-4454381" [ 1083.354546] env[65385]: _type = "Task" [ 1083.354546] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.355360] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d920c15-ca08-49b0-9c9d-79d5b52789bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.371174] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.376686] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.376947] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.378208] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9b74dee-5b1b-4e76-81f5-8279401639df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.385193] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454378, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.813017} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.385580] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1083.385818] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.386160] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6e97ef4-2792-4758-9b55-42c3f23076c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.392948] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1083.392948] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dbde19-070a-7ade-56e7-32f4930840e6" [ 1083.392948] env[65385]: _type = "Task" [ 1083.392948] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.404386] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1083.404386] env[65385]: value = "task-4454382" [ 1083.404386] env[65385]: _type = "Task" [ 1083.404386] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.416638] env[65385]: DEBUG oslo_vmware.api [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454379, 'name': ReconfigVM_Task, 'duration_secs': 0.851489} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.417528] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dbde19-070a-7ade-56e7-32f4930840e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.418353] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871167', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'name': 'volume-1b184555-e286-4867-9a99-c29012c47f92', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5470afe7-6dd0-4778-9b83-2956ede04b7f', 'attached_at': '', 'detached_at': '', 'volume_id': '1b184555-e286-4867-9a99-c29012c47f92', 'serial': '1b184555-e286-4867-9a99-c29012c47f92'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1083.425415] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454382, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.673653] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.680170] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.489s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.680525] env[65385]: DEBUG nova.objects.instance [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'resources' on Instance uuid 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.681740] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071876} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.682396] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1083.683363] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d12aeda-9bef-437f-b37b-4490655bbd40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.709742] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.712544] env[65385]: INFO nova.scheduler.client.report [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted allocations for instance 77d718c8-4959-483a-8717-c00f2be1d0bb [ 1083.713558] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e78e34b-db83-41ad-a35e-98a63d1d5221 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.741193] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1083.741193] env[65385]: value = "task-4454383" [ 1083.741193] env[65385]: _type = "Task" [ 1083.741193] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.753351] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454383, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.813176] env[65385]: DEBUG nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1083.847185] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1083.847501] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1083.847698] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1083.847952] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1083.848431] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1083.848598] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1083.848859] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1083.849033] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1083.849210] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1083.850093] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1083.850345] env[65385]: DEBUG nova.virt.hardware [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1083.851330] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2477eb55-d950-4835-9f44-d31955c21093 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.864280] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0917bbf-710e-4431-8a02-d8f4759923f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.878044] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454381, 'name': PowerOffVM_Task, 'duration_secs': 0.408974} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.886342] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1083.886577] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance '29a449d0-f62d-4430-96ab-9ac43e5a7ad2' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.909495] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dbde19-070a-7ade-56e7-32f4930840e6, 'name': SearchDatastore_Task, 'duration_secs': 0.022452} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.913617] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-240d47c9-c5f8-4cde-8425-af4c40eba9af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.922029] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454382, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074651} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.923563] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1083.924025] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1083.924025] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5271ec0f-06ae-d31e-5ccc-e6e92023bf45" [ 1083.924025] env[65385]: _type = "Task" [ 1083.924025] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.924847] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35fb655-11f1-40b2-b91c-9868e0a5d1cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.952885] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.957195] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e3f804e-18ba-4bbc-94ed-e0f7b6244166 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.971948] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5271ec0f-06ae-d31e-5ccc-e6e92023bf45, 'name': SearchDatastore_Task, 'duration_secs': 0.010369} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.972976] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.973256] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 805528b8-96db-4bb8-ba05-2b6332c54624/805528b8-96db-4bb8-ba05-2b6332c54624.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1083.974093] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54d9ded3-266e-41ef-ab23-aaac9443685f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.980195] env[65385]: DEBUG nova.objects.instance [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'flavor' on Instance uuid 5470afe7-6dd0-4778-9b83-2956ede04b7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.983551] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1083.983551] env[65385]: value = "task-4454384" [ 1083.983551] env[65385]: _type = "Task" [ 1083.983551] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.985324] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1083.985324] env[65385]: value = "task-4454385" [ 1083.985324] env[65385]: _type = "Task" [ 1083.985324] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.004767] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454385, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.005051] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.029808] env[65385]: DEBUG nova.compute.manager [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1084.029808] env[65385]: DEBUG nova.compute.manager [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing instance network info cache due to event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1084.030133] env[65385]: DEBUG oslo_concurrency.lockutils [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.030231] env[65385]: DEBUG oslo_concurrency.lockutils [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.031211] env[65385]: DEBUG nova.network.neutron [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1084.237804] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ed9cac0b-ba42-4a70-b3cc-a82cfac4221e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "77d718c8-4959-483a-8717-c00f2be1d0bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.874s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.258574] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454383, 'name': ReconfigVM_Task, 'duration_secs': 0.421982} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.258574] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.258777] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14721495-c34b-489b-a904-c7ac896833c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.273062] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1084.273062] env[65385]: value = "task-4454386" [ 1084.273062] env[65385]: _type = "Task" [ 1084.273062] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.288443] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454386, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.385673] env[65385]: DEBUG nova.compute.manager [req-b5f9552c-e66a-4ba4-8f1c-a95e0bc4caa7 req-4d2086f0-96da-47a7-80f5-b2792f354bfb service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received event network-vif-plugged-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1084.385673] env[65385]: DEBUG oslo_concurrency.lockutils [req-b5f9552c-e66a-4ba4-8f1c-a95e0bc4caa7 req-4d2086f0-96da-47a7-80f5-b2792f354bfb service nova] Acquiring lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.386015] env[65385]: DEBUG oslo_concurrency.lockutils [req-b5f9552c-e66a-4ba4-8f1c-a95e0bc4caa7 req-4d2086f0-96da-47a7-80f5-b2792f354bfb service nova] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.386288] env[65385]: DEBUG oslo_concurrency.lockutils [req-b5f9552c-e66a-4ba4-8f1c-a95e0bc4caa7 req-4d2086f0-96da-47a7-80f5-b2792f354bfb service nova] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.386531] env[65385]: DEBUG nova.compute.manager [req-b5f9552c-e66a-4ba4-8f1c-a95e0bc4caa7 req-4d2086f0-96da-47a7-80f5-b2792f354bfb service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] No waiting events found dispatching network-vif-plugged-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1084.387138] env[65385]: WARNING nova.compute.manager [req-b5f9552c-e66a-4ba4-8f1c-a95e0bc4caa7 req-4d2086f0-96da-47a7-80f5-b2792f354bfb service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received unexpected event network-vif-plugged-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 for instance with vm_state building and task_state spawning. [ 1084.395312] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1084.395622] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1084.395824] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1084.396043] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1084.396241] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1084.396453] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1084.396707] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.396885] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1084.397168] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1084.397340] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1084.397551] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1084.407781] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df8d9e7c-1687-41a4-9989-80fc0a000fd8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.430401] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1084.430401] env[65385]: value = "task-4454387" [ 1084.430401] env[65385]: _type = "Task" [ 1084.430401] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.444663] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454387, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.449442] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58afa8a-eaae-438d-a4aa-2fca58751a2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.459750] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fcb171-1cf1-4526-92a9-2eff5d4c8cf9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.500646] env[65385]: DEBUG nova.network.neutron [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Successfully updated port: bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1084.510654] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83bf8af-bb1a-4a79-ae4c-0e87e977f1fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.526377] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454385, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509644} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.526636] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.528482] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71efd5b7-0197-409a-bd2c-7d125f7eed6d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.533188] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 805528b8-96db-4bb8-ba05-2b6332c54624/805528b8-96db-4bb8-ba05-2b6332c54624.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1084.533917] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1084.534012] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08a6c529-a872-4149-9772-4723592fd50a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.536471] env[65385]: WARNING neutronclient.v2_0.client [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.537147] env[65385]: WARNING openstack [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.537538] env[65385]: WARNING openstack [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.557964] env[65385]: DEBUG nova.compute.provider_tree [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.561568] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1084.561568] env[65385]: value = "task-4454388" [ 1084.561568] env[65385]: _type = "Task" [ 1084.561568] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.573907] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454388, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.709127] env[65385]: WARNING openstack [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.709517] env[65385]: WARNING openstack [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.773121] env[65385]: WARNING neutronclient.v2_0.client [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.773879] env[65385]: WARNING openstack [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.774308] env[65385]: WARNING openstack [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.791812] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454386, 'name': Rename_Task, 'duration_secs': 0.24348} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.792086] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.792380] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dff87a82-b58a-490e-9850-c3339700a3b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.799506] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1084.799506] env[65385]: value = "task-4454389" [ 1084.799506] env[65385]: _type = "Task" [ 1084.799506] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.808295] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.860880] env[65385]: DEBUG nova.network.neutron [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updated VIF entry in instance network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1084.861289] env[65385]: DEBUG nova.network.neutron [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1084.941407] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454387, 'name': ReconfigVM_Task, 'duration_secs': 0.248036} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.941925] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance '29a449d0-f62d-4430-96ab-9ac43e5a7ad2' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1085.012166] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454384, 'name': ReconfigVM_Task, 'duration_secs': 0.647731} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.012456] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c/5ae074b9-4e82-4345-9bc1-ab8d7c930f1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.013147] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b63428b-21ff-4151-aff5-b6c2e98e7cc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.015626] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.015787] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquired lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1085.015960] env[65385]: DEBUG nova.network.neutron [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1085.017242] env[65385]: DEBUG oslo_concurrency.lockutils [None req-60c2bfba-1a18-47a5-a68e-de3ecf8ab8dd tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.388s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.024495] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1085.024495] env[65385]: value = "task-4454390" [ 1085.024495] env[65385]: _type = "Task" [ 1085.024495] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.036349] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454390, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.064126] env[65385]: DEBUG nova.scheduler.client.report [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1085.078307] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454388, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079006} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.078691] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.079572] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022521e4-7b9f-4d4c-8c0f-bd92dca6ea4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.104469] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 805528b8-96db-4bb8-ba05-2b6332c54624/805528b8-96db-4bb8-ba05-2b6332c54624.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.106648] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec5b3e39-16b8-41dc-8e18-be30dd44e43e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.125623] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.126131] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.134878] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1085.134878] env[65385]: value = "task-4454391" [ 1085.134878] env[65385]: _type = "Task" [ 1085.134878] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.145054] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454391, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.315055] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454389, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.365308] env[65385]: DEBUG oslo_concurrency.lockutils [req-3422bd22-e042-4826-9121-3ea4bad339f7 req-2fbed4ea-d003-48fc-9adc-c481fd8f96ba service nova] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.448398] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1085.448632] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1085.448783] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1085.448959] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1085.449112] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1085.449259] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1085.449458] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.449608] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1085.449768] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1085.449925] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1085.450111] env[65385]: DEBUG nova.virt.hardware [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1085.455571] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfiguring VM instance instance-00000050 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1085.455891] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4126effa-88d2-4f0b-9c14-63d4f5ed2753 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.475523] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1085.475523] env[65385]: value = "task-4454392" [ 1085.475523] env[65385]: _type = "Task" [ 1085.475523] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.484957] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454392, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.520160] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.520741] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.540555] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454390, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.568943] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.571897] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.466s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.572207] env[65385]: DEBUG nova.objects.instance [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'resources' on Instance uuid 04ebe83c-cab5-45e1-9f5f-018481f63912 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.583727] env[65385]: DEBUG nova.network.neutron [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1085.593681] env[65385]: INFO nova.scheduler.client.report [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted allocations for instance 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7 [ 1085.618955] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.619497] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.629766] env[65385]: DEBUG nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1085.645285] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454391, 'name': ReconfigVM_Task, 'duration_secs': 0.429543} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.645643] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 805528b8-96db-4bb8-ba05-2b6332c54624/805528b8-96db-4bb8-ba05-2b6332c54624.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.646396] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bbdb9198-c711-4860-bcdd-87a15f0a4f5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.654594] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1085.654594] env[65385]: value = "task-4454393" [ 1085.654594] env[65385]: _type = "Task" [ 1085.654594] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.670381] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454393, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.680529] env[65385]: WARNING neutronclient.v2_0.client [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1085.681398] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.681828] env[65385]: WARNING openstack [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.767319] env[65385]: DEBUG nova.network.neutron [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [{"id": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "address": "fa:16:3e:49:27:54", "network": {"id": "380eb239-46a9-4bb2-be3e-66bb113acf76", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1139908934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0a282a49e5f44dc781471d9c2103ca8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38f5da-81", "ovs_interfaceid": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1085.812249] env[65385]: DEBUG oslo_vmware.api [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454389, 'name': PowerOnVM_Task, 'duration_secs': 0.638369} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.812559] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.812730] env[65385]: INFO nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Took 8.64 seconds to spawn the instance on the hypervisor. [ 1085.812907] env[65385]: DEBUG nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1085.814136] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351b6656-7c90-44a0-acec-7696d938c8c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.986182] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454392, 'name': ReconfigVM_Task, 'duration_secs': 0.320229} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.986470] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfigured VM instance instance-00000050 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1085.987322] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd932129-667a-4fc0-9d52-99935aa6d3c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.990210] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.990525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.990732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "5470afe7-6dd0-4778-9b83-2956ede04b7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.991016] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.991291] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.994053] env[65385]: INFO nova.compute.manager [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Terminating instance [ 1086.025450] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.027012] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76ecb86d-ce20-448e-a2a6-02510bccf557 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.053703] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454390, 'name': Rename_Task, 'duration_secs': 0.784228} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.055565] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1086.055930] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1086.055930] env[65385]: value = "task-4454394" [ 1086.055930] env[65385]: _type = "Task" [ 1086.055930] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.056351] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d6ed108-3c10-491b-b4af-bdb31334511f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.068268] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1086.068268] env[65385]: value = "task-4454395" [ 1086.068268] env[65385]: _type = "Task" [ 1086.068268] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.071859] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454394, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.085926] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.101203] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3077c864-d5fc-4f43-91d0-577a6b935118 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.855s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.154655] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.164946] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454393, 'name': Rename_Task, 'duration_secs': 0.443211} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.167775] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1086.168471] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4163face-7487-4bdd-94c5-9465d84ebed8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.176666] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1086.176666] env[65385]: value = "task-4454396" [ 1086.176666] env[65385]: _type = "Task" [ 1086.176666] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.190292] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454396, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.270444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Releasing lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1086.271137] env[65385]: DEBUG nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Instance network_info: |[{"id": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "address": "fa:16:3e:49:27:54", "network": {"id": "380eb239-46a9-4bb2-be3e-66bb113acf76", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1139908934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0a282a49e5f44dc781471d9c2103ca8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38f5da-81", "ovs_interfaceid": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1086.272100] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:27:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.282123] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Creating folder: Project (0a282a49e5f44dc781471d9c2103ca8f). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1086.285956] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-343b8124-b0e4-4f85-aa2e-fc84ea32893d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.300704] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Created folder: Project (0a282a49e5f44dc781471d9c2103ca8f) in parent group-v870881. [ 1086.300969] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Creating folder: Instances. Parent ref: group-v871172. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1086.301244] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b8464cd8-d10b-49e7-b2ce-3abc9a06e30d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.308103] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c2eced-693f-4c4e-8396-c06f9a92abd7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.316168] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Created folder: Instances in parent group-v871172. [ 1086.316645] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1086.316861] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1086.317190] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f65a5ef1-77b1-4416-a44e-84fbd2993e05 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.340221] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fb696b-42b9-4b78-b6bd-4d66ece30069 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.347709] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.347709] env[65385]: value = "task-4454399" [ 1086.347709] env[65385]: _type = "Task" [ 1086.347709] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.348494] env[65385]: INFO nova.compute.manager [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Took 14.19 seconds to build instance. [ 1086.384689] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e75349d-8fd3-4ad5-a626-ced3b93182ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.391703] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454399, 'name': CreateVM_Task} progress is 15%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.399239] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99bb2cdf-040e-4e06-80d8-49b16c2bcde5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.417366] env[65385]: DEBUG nova.compute.provider_tree [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.540170] env[65385]: DEBUG nova.compute.manager [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1086.540557] env[65385]: DEBUG nova.compute.manager [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing instance network info cache due to event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1086.541491] env[65385]: DEBUG oslo_concurrency.lockutils [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Acquiring lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.541682] env[65385]: DEBUG oslo_concurrency.lockutils [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Acquired lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.542042] env[65385]: DEBUG nova.network.neutron [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1086.544569] env[65385]: DEBUG nova.compute.manager [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1086.544782] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1086.545892] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a5fc38-4b17-41df-9d42-6751bed6a633 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.556946] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.558178] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74d3a466-1e64-4344-8bf6-77984aab0cbc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.568772] env[65385]: DEBUG oslo_vmware.api [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1086.568772] env[65385]: value = "task-4454400" [ 1086.568772] env[65385]: _type = "Task" [ 1086.568772] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.573267] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454394, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.590076] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454395, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.594227] env[65385]: DEBUG oslo_vmware.api [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.691792] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454396, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.807349] env[65385]: DEBUG nova.compute.manager [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Received event network-changed-92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1086.808074] env[65385]: DEBUG nova.compute.manager [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Refreshing instance network info cache due to event network-changed-92c45f41-92b4-40c3-8516-30189d29e33d. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1086.808570] env[65385]: DEBUG oslo_concurrency.lockutils [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Acquiring lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.808815] env[65385]: DEBUG oslo_concurrency.lockutils [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Acquired lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.809093] env[65385]: DEBUG nova.network.neutron [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Refreshing network info cache for port 92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1086.861405] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454399, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.883862] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bc6ad4f6-d34e-4dc4-a324-303175f08ad0 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.739s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.920926] env[65385]: DEBUG nova.scheduler.client.report [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1087.046246] env[65385]: WARNING neutronclient.v2_0.client [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.047468] env[65385]: WARNING openstack [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.047826] env[65385]: WARNING openstack [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.071653] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454394, 'name': ReconfigVM_Task, 'duration_secs': 0.567163} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.071959] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.072255] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance '29a449d0-f62d-4430-96ab-9ac43e5a7ad2' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.089348] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454395, 'name': PowerOnVM_Task, 'duration_secs': 0.94691} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.092561] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.092778] env[65385]: DEBUG nova.compute.manager [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1087.093083] env[65385]: DEBUG oslo_vmware.api [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454400, 'name': PowerOffVM_Task, 'duration_secs': 0.432421} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.093844] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffdf67c-677f-4105-990b-39f192623cc9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.096693] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.096873] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1087.097157] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c21f861c-2d61-472f-ba56-7bb71348623a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.186049] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1087.186415] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1087.186620] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleting the datastore file [datastore1] 5470afe7-6dd0-4778-9b83-2956ede04b7f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.190012] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-951f67a5-8317-40bb-b2a3-6f7a23d5543b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.192170] env[65385]: DEBUG oslo_vmware.api [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454396, 'name': PowerOnVM_Task, 'duration_secs': 0.636913} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.192810] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.192810] env[65385]: INFO nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Took 8.21 seconds to spawn the instance on the hypervisor. [ 1087.192999] env[65385]: DEBUG nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1087.194357] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef92128-ff0a-4472-82e2-f8d05f722c1c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.198319] env[65385]: WARNING openstack [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.198706] env[65385]: WARNING openstack [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.209652] env[65385]: DEBUG oslo_vmware.api [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1087.209652] env[65385]: value = "task-4454402" [ 1087.209652] env[65385]: _type = "Task" [ 1087.209652] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.221502] env[65385]: DEBUG oslo_vmware.api [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454402, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.264341] env[65385]: WARNING neutronclient.v2_0.client [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.265093] env[65385]: WARNING openstack [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.265494] env[65385]: WARNING openstack [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.314973] env[65385]: WARNING neutronclient.v2_0.client [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.315793] env[65385]: WARNING openstack [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.317091] env[65385]: WARNING openstack [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.363921] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454399, 'name': CreateVM_Task, 'duration_secs': 0.610009} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.366401] env[65385]: DEBUG nova.network.neutron [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updated VIF entry in instance network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1087.366781] env[65385]: DEBUG nova.network.neutron [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [{"id": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "address": "fa:16:3e:49:27:54", "network": {"id": "380eb239-46a9-4bb2-be3e-66bb113acf76", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1139908934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0a282a49e5f44dc781471d9c2103ca8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38f5da-81", "ovs_interfaceid": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1087.370306] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1087.370539] env[65385]: WARNING neutronclient.v2_0.client [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.370894] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.371234] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.371877] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1087.371989] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d5b9fda-c168-4140-8773-4b1bef9c9f47 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.379660] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1087.379660] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d5008d-7875-e9c4-6a2c-949ced9162bd" [ 1087.379660] env[65385]: _type = "Task" [ 1087.379660] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.389318] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d5008d-7875-e9c4-6a2c-949ced9162bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.428791] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1087.431732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.277s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.434876] env[65385]: INFO nova.compute.claims [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.458821] env[65385]: INFO nova.scheduler.client.report [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted allocations for instance 04ebe83c-cab5-45e1-9f5f-018481f63912 [ 1087.586721] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e062b42-47d1-4609-8010-9cea4084b322 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.611957] env[65385]: INFO nova.compute.manager [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] bringing vm to original state: 'stopped' [ 1087.616067] env[65385]: WARNING openstack [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.617291] env[65385]: WARNING openstack [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.624757] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0104a004-632c-46fa-ac3b-b586f761e2ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.654582] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance '29a449d0-f62d-4430-96ab-9ac43e5a7ad2' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.724720] env[65385]: DEBUG oslo_vmware.api [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454402, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160886} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.725373] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1087.726616] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1087.726616] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1087.726616] env[65385]: INFO nova.compute.manager [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1087.726731] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1087.730876] env[65385]: DEBUG nova.compute.manager [-] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1087.730876] env[65385]: DEBUG nova.network.neutron [-] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1087.730876] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.730876] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.730876] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.738824] env[65385]: INFO nova.compute.manager [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Took 15.17 seconds to build instance. [ 1087.741366] env[65385]: WARNING neutronclient.v2_0.client [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.742052] env[65385]: WARNING openstack [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.742539] env[65385]: WARNING openstack [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.805178] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.841046] env[65385]: DEBUG nova.network.neutron [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updated VIF entry in instance network info cache for port 92c45f41-92b4-40c3-8516-30189d29e33d. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1087.841427] env[65385]: DEBUG nova.network.neutron [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1087.871051] env[65385]: DEBUG oslo_concurrency.lockutils [req-3ea80b25-983c-4625-a783-8e254e6b8f4a req-0a6716e3-b4ca-4b4d-b4f9-4dc249ef9cc9 service nova] Releasing lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.890391] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d5008d-7875-e9c4-6a2c-949ced9162bd, 'name': SearchDatastore_Task, 'duration_secs': 0.011469} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.890692] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.890972] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.891261] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.891408] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.891605] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.892245] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aca6040b-f9b7-419e-a4fa-a75ed2e034cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.905802] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.905998] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1087.906861] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c4c1772-e605-4d26-ba66-378c16922023 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.914774] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1087.914774] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d9f95c-9f85-f619-77ba-5015000398fb" [ 1087.914774] env[65385]: _type = "Task" [ 1087.914774] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.925960] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d9f95c-9f85-f619-77ba-5015000398fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.966213] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b71c177f-7b56-433b-937f-071f6c3cb2fa tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "04ebe83c-cab5-45e1-9f5f-018481f63912" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.277s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.241054] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b809e16-a1ca-4c70-88b7-018e0412a547 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.689s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.344390] env[65385]: DEBUG oslo_concurrency.lockutils [req-7cf1f37a-78a7-49b8-88bc-e0565043eb0d req-114dc5e5-37e2-4c42-9ecf-2c5291fc16a1 service nova] Releasing lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.430027] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d9f95c-9f85-f619-77ba-5015000398fb, 'name': SearchDatastore_Task, 'duration_secs': 0.039536} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.431866] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9ba3371-4128-433a-a309-3516f63647ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.438916] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1088.438916] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52880542-b4b3-1a2c-3800-60a8c4ef918e" [ 1088.438916] env[65385]: _type = "Task" [ 1088.438916] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.453435] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52880542-b4b3-1a2c-3800-60a8c4ef918e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.612600] env[65385]: DEBUG nova.compute.manager [req-fe5a2381-12b3-4f35-9463-824d951f5db5 req-5b6edd1a-9273-4456-be58-febb873beb46 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Received event network-vif-deleted-2ea48afc-dea7-435d-996b-e1f73fcec4bf {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1088.612600] env[65385]: INFO nova.compute.manager [req-fe5a2381-12b3-4f35-9463-824d951f5db5 req-5b6edd1a-9273-4456-be58-febb873beb46 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Neutron deleted interface 2ea48afc-dea7-435d-996b-e1f73fcec4bf; detaching it from the instance and deleting it from the info cache [ 1088.612600] env[65385]: DEBUG nova.network.neutron [req-fe5a2381-12b3-4f35-9463-824d951f5db5 req-5b6edd1a-9273-4456-be58-febb873beb46 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.633218] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.633357] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.633527] env[65385]: DEBUG nova.compute.manager [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1088.637760] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9430b2-f14f-4473-bdd6-6a8d7ab2e986 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.648576] env[65385]: DEBUG nova.compute.manager [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1088.682768] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54de309f-7823-42f4-9f37-407e5b7f9e01 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.692710] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8466627-8077-415e-ba69-5f56a1748493 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.727040] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b845f31d-6761-4aae-b80a-7dfa95b26bcc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.736707] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04c0a89-d249-4fd3-9a1a-f02ccaf1cba4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.752650] env[65385]: DEBUG nova.compute.provider_tree [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.789461] env[65385]: DEBUG nova.network.neutron [-] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.916302] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.917055] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.954052] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52880542-b4b3-1a2c-3800-60a8c4ef918e, 'name': SearchDatastore_Task, 'duration_secs': 0.04577} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.954228] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.954483] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/f0f981dd-2c6b-4b3a-b62f-4b295fe3457c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1088.954750] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1de027d3-45c3-45bb-bb93-c76ef80fe0f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.962979] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1088.962979] env[65385]: value = "task-4454403" [ 1088.962979] env[65385]: _type = "Task" [ 1088.962979] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.972625] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454403, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.013655] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.014040] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1089.115509] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d3e1061-5c95-446f-9184-f4bb11cc7cd6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.126209] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ee9a5f-0de8-4b88-850b-1071c282a3e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.162624] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1089.163704] env[65385]: DEBUG nova.compute.manager [req-fe5a2381-12b3-4f35-9463-824d951f5db5 req-5b6edd1a-9273-4456-be58-febb873beb46 service nova] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Detach interface failed, port_id=2ea48afc-dea7-435d-996b-e1f73fcec4bf, reason: Instance 5470afe7-6dd0-4778-9b83-2956ede04b7f could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1089.163971] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ce18848-7e82-4abd-ac91-888157cdf36d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.171326] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1089.171326] env[65385]: value = "task-4454404" [ 1089.171326] env[65385]: _type = "Task" [ 1089.171326] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.181296] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454404, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.256051] env[65385]: DEBUG nova.scheduler.client.report [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1089.292352] env[65385]: INFO nova.compute.manager [-] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Took 1.56 seconds to deallocate network for instance. [ 1089.317615] env[65385]: WARNING neutronclient.v2_0.client [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1089.317615] env[65385]: WARNING neutronclient.v2_0.client [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1089.369065] env[65385]: DEBUG nova.network.neutron [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Port ace973a7-f2da-4905-99cb-e72ce9ef3f18 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1089.418939] env[65385]: DEBUG nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1089.475099] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454403, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.517246] env[65385]: DEBUG nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1089.683336] env[65385]: DEBUG oslo_vmware.api [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454404, 'name': PowerOffVM_Task, 'duration_secs': 0.323891} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.683336] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1089.683519] env[65385]: DEBUG nova.compute.manager [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1089.684342] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb33b1f-5303-45cd-8fb9-8e5d69e84a3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.762057] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.762491] env[65385]: DEBUG nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1089.802070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.802070] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1089.802275] env[65385]: DEBUG nova.objects.instance [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'resources' on Instance uuid 5470afe7-6dd0-4778-9b83-2956ede04b7f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.942861] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.964122] env[65385]: DEBUG nova.compute.manager [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1089.977143] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454403, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571873} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.977444] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/f0f981dd-2c6b-4b3a-b62f-4b295fe3457c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1089.977650] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.978235] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af6adeac-be96-4c95-84ab-f817ff12b032 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.986344] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1089.986344] env[65385]: value = "task-4454405" [ 1089.986344] env[65385]: _type = "Task" [ 1089.986344] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.997110] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454405, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.037854] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.197382] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.564s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.267997] env[65385]: DEBUG nova.compute.utils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1090.270739] env[65385]: DEBUG nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1090.270739] env[65385]: DEBUG nova.network.neutron [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1090.270739] env[65385]: WARNING neutronclient.v2_0.client [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1090.271089] env[65385]: WARNING neutronclient.v2_0.client [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1090.271834] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1090.272434] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1090.319839] env[65385]: DEBUG nova.policy [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1090.399369] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.399611] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.399781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.484661] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.497679] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454405, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150436} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.500372] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1090.501627] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc926b2-0f9a-4496-8f67-37b6c5d2ac60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.527727] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/f0f981dd-2c6b-4b3a-b62f-4b295fe3457c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1090.531358] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd4b2e93-823e-484c-a4a8-3dd37a1985dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.563658] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1090.563658] env[65385]: value = "task-4454406" [ 1090.563658] env[65385]: _type = "Task" [ 1090.563658] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.570396] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe9bfba-d1d5-4509-9d15-638efa014454 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.576751] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454406, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.582143] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed894ee-65e3-420b-9483-a22b6e9a1ba0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.614174] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea3dff5-8dbc-477e-af79-89efb0bfdab2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.622738] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf68be0-ae7e-43a2-a129-dca781ae0fd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.640535] env[65385]: DEBUG nova.compute.provider_tree [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.647107] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.647365] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.647637] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.647819] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.648246] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.650405] env[65385]: INFO nova.compute.manager [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Terminating instance [ 1090.670378] env[65385]: DEBUG nova.network.neutron [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Successfully created port: 6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1090.705042] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.782422] env[65385]: DEBUG nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1091.074634] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454406, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.144186] env[65385]: DEBUG nova.scheduler.client.report [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1091.155702] env[65385]: DEBUG nova.compute.manager [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1091.155974] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.156876] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853c4ff1-8760-40c3-a31e-d7a69e8647c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.166314] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.167256] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3ba52be-7240-4d9d-b83a-9f902a956893 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.248947] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.249265] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.249496] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleting the datastore file [datastore2] 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.249824] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-283d6ff2-3b66-49ce-8bb5-079de5674440 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.258269] env[65385]: DEBUG oslo_vmware.api [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1091.258269] env[65385]: value = "task-4454408" [ 1091.258269] env[65385]: _type = "Task" [ 1091.258269] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.267732] env[65385]: DEBUG oslo_vmware.api [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454408, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.406752] env[65385]: WARNING neutronclient.v2_0.client [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.440255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.440730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.440988] env[65385]: DEBUG nova.network.neutron [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1091.575920] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454406, 'name': ReconfigVM_Task, 'duration_secs': 0.892366} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.576234] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Reconfigured VM instance instance-00000068 to attach disk [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/f0f981dd-2c6b-4b3a-b62f-4b295fe3457c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1091.576923] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1b46b6c9-83ca-4dad-a352-e822730b55f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.585633] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1091.585633] env[65385]: value = "task-4454409" [ 1091.585633] env[65385]: _type = "Task" [ 1091.585633] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.595613] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454409, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.649734] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.652286] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.709s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.653893] env[65385]: INFO nova.compute.claims [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1091.674201] env[65385]: INFO nova.scheduler.client.report [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted allocations for instance 5470afe7-6dd0-4778-9b83-2956ede04b7f [ 1091.769283] env[65385]: DEBUG oslo_vmware.api [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454408, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182842} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.769590] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.769778] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1091.769955] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1091.770145] env[65385]: INFO nova.compute.manager [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1091.770387] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1091.770585] env[65385]: DEBUG nova.compute.manager [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1091.770682] env[65385]: DEBUG nova.network.neutron [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1091.770930] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.771477] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.771794] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.793331] env[65385]: DEBUG nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1091.810532] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.820393] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1091.820393] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1091.820558] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1091.820710] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1091.820853] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1091.820996] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1091.821220] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.821376] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1091.821535] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1091.821693] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1091.821863] env[65385]: DEBUG nova.virt.hardware [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1091.822723] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1deff4b5-e7be-4419-af85-c034c68e4c41 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.831664] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c998230d-6ebf-44c8-ae05-4282b64ea405 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.945113] env[65385]: WARNING neutronclient.v2_0.client [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.945926] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.946353] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.064456] env[65385]: DEBUG nova.compute.manager [req-ab9a5017-792a-43b5-994f-46e3565064f9 req-64f1fd55-4234-4121-93d7-22b8e555154e service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Received event network-vif-deleted-d33bca4a-a999-48ad-8138-b0f8d5b6ee23 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1092.065656] env[65385]: INFO nova.compute.manager [req-ab9a5017-792a-43b5-994f-46e3565064f9 req-64f1fd55-4234-4121-93d7-22b8e555154e service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Neutron deleted interface d33bca4a-a999-48ad-8138-b0f8d5b6ee23; detaching it from the instance and deleting it from the info cache [ 1092.065656] env[65385]: DEBUG nova.network.neutron [req-ab9a5017-792a-43b5-994f-46e3565064f9 req-64f1fd55-4234-4121-93d7-22b8e555154e service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.086038] env[65385]: DEBUG nova.compute.manager [req-3967529b-034d-4aad-b7f0-0cab8abedfc4 req-cf662cd8-7fa3-4a6e-aba8-8e24ae8855d0 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-vif-plugged-6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1092.086298] env[65385]: DEBUG oslo_concurrency.lockutils [req-3967529b-034d-4aad-b7f0-0cab8abedfc4 req-cf662cd8-7fa3-4a6e-aba8-8e24ae8855d0 service nova] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.086575] env[65385]: DEBUG oslo_concurrency.lockutils [req-3967529b-034d-4aad-b7f0-0cab8abedfc4 req-cf662cd8-7fa3-4a6e-aba8-8e24ae8855d0 service nova] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.086753] env[65385]: DEBUG oslo_concurrency.lockutils [req-3967529b-034d-4aad-b7f0-0cab8abedfc4 req-cf662cd8-7fa3-4a6e-aba8-8e24ae8855d0 service nova] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.086922] env[65385]: DEBUG nova.compute.manager [req-3967529b-034d-4aad-b7f0-0cab8abedfc4 req-cf662cd8-7fa3-4a6e-aba8-8e24ae8855d0 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] No waiting events found dispatching network-vif-plugged-6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1092.087097] env[65385]: WARNING nova.compute.manager [req-3967529b-034d-4aad-b7f0-0cab8abedfc4 req-cf662cd8-7fa3-4a6e-aba8-8e24ae8855d0 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received unexpected event network-vif-plugged-6e6c199b-75f0-473b-90b5-0daaf633ba14 for instance with vm_state building and task_state spawning. [ 1092.097962] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454409, 'name': Rename_Task, 'duration_secs': 0.1634} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.100741] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1092.101227] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2518ed02-8fb4-4e80-bf25-497a839ccdae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.109515] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1092.109515] env[65385]: value = "task-4454410" [ 1092.109515] env[65385]: _type = "Task" [ 1092.109515] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.118380] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.185613] env[65385]: DEBUG nova.network.neutron [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Successfully updated port: 6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1092.195605] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fd17ad21-4202-440d-9732-ae3f79f0ec06 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "5470afe7-6dd0-4778-9b83-2956ede04b7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.205s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.215386] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.215866] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.281643] env[65385]: WARNING neutronclient.v2_0.client [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.282374] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.282764] env[65385]: WARNING openstack [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.378689] env[65385]: DEBUG nova.network.neutron [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.536753] env[65385]: DEBUG nova.network.neutron [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.569860] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1fd5e56-e106-4763-be40-e13161475715 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.581583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e789db-fe9b-4c09-91b7-fedc4ba1e051 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.620622] env[65385]: DEBUG nova.compute.manager [req-ab9a5017-792a-43b5-994f-46e3565064f9 req-64f1fd55-4234-4121-93d7-22b8e555154e service nova] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Detach interface failed, port_id=d33bca4a-a999-48ad-8138-b0f8d5b6ee23, reason: Instance 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1092.632366] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454410, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.688828] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.689046] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1092.689275] env[65385]: DEBUG nova.network.neutron [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1092.882691] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.899376] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284b7fd6-8a3b-4d6a-9554-868c48473390 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.908554] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c49b17b-dad3-4369-ad5d-5b935a7074d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.941815] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795520f2-f3f2-432c-9248-e17ddfc890cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.951442] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dec64fd-7049-451d-b54e-52a568b0ca77 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.968173] env[65385]: DEBUG nova.compute.provider_tree [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.039593] env[65385]: INFO nova.compute.manager [-] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Took 1.27 seconds to deallocate network for instance. [ 1093.130796] env[65385]: DEBUG oslo_vmware.api [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454410, 'name': PowerOnVM_Task, 'duration_secs': 0.666652} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.131105] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1093.131338] env[65385]: INFO nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Took 9.32 seconds to spawn the instance on the hypervisor. [ 1093.131529] env[65385]: DEBUG nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1093.132432] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df043ee9-e569-4a40-b6a4-7fe41d7a2210 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.192792] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.192891] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.227878] env[65385]: DEBUG nova.network.neutron [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1093.248728] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.248728] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.334155] env[65385]: WARNING neutronclient.v2_0.client [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.334876] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.335271] env[65385]: WARNING openstack [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.392965] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1acd92d9-810e-4d8b-8855-57111bcc1d21 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.404983] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aece1e3-8747-46ac-8cf2-23c467ee1424 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.472099] env[65385]: DEBUG nova.scheduler.client.report [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1093.491422] env[65385]: DEBUG nova.network.neutron [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1093.547274] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.651801] env[65385]: INFO nova.compute.manager [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Took 20.07 seconds to build instance. [ 1093.977731] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.978276] env[65385]: DEBUG nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1093.981072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.943s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.982562] env[65385]: INFO nova.compute.claims [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1093.995033] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1093.995228] env[65385]: DEBUG nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Instance network_info: |[{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1093.995741] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:36:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e6c199b-75f0-473b-90b5-0daaf633ba14', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1094.003829] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1094.004145] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1094.004424] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b0cf3f4-ed55-49a4-bbf8-12fdf27b8566 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.030148] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1094.030148] env[65385]: value = "task-4454411" [ 1094.030148] env[65385]: _type = "Task" [ 1094.030148] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.039819] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454411, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.119790] env[65385]: DEBUG nova.compute.manager [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1094.120184] env[65385]: DEBUG nova.compute.manager [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing instance network info cache due to event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1094.120475] env[65385]: DEBUG oslo_concurrency.lockutils [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.120833] env[65385]: DEBUG oslo_concurrency.lockutils [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.121057] env[65385]: DEBUG nova.network.neutron [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1094.154417] env[65385]: DEBUG oslo_concurrency.lockutils [None req-6909f67a-6115-47e0-a638-d9ed1dc99137 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.585s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.248596] env[65385]: INFO nova.compute.manager [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Rescuing [ 1094.249259] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.249259] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquired lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.249259] env[65385]: DEBUG nova.network.neutron [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1094.488063] env[65385]: DEBUG nova.compute.utils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1094.490976] env[65385]: DEBUG nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1094.491209] env[65385]: DEBUG nova.network.neutron [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1094.491556] env[65385]: WARNING neutronclient.v2_0.client [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.492160] env[65385]: WARNING neutronclient.v2_0.client [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.492517] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.494163] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.538245] env[65385]: DEBUG nova.policy [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77dc7acbf0324ac88b7092803faaa384', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b27c092a0fd345ea953956469c21f8bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1094.542543] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cebe051-f738-45d1-87a6-8581774003e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.549030] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454411, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.570648] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb99d4e5-7d74-4f67-9249-8592491e98eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.580287] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance '29a449d0-f62d-4430-96ab-9ac43e5a7ad2' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1094.624846] env[65385]: WARNING neutronclient.v2_0.client [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.625374] env[65385]: WARNING openstack [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.625747] env[65385]: WARNING openstack [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.751078] env[65385]: WARNING openstack [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.751523] env[65385]: WARNING openstack [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.760699] env[65385]: WARNING neutronclient.v2_0.client [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.761622] env[65385]: WARNING openstack [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.762106] env[65385]: WARNING openstack [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.828134] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.828134] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.895740] env[65385]: WARNING neutronclient.v2_0.client [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.896546] env[65385]: WARNING openstack [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.896969] env[65385]: WARNING openstack [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.926774] env[65385]: DEBUG nova.network.neutron [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Successfully created port: cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1094.989854] env[65385]: WARNING openstack [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.990293] env[65385]: WARNING openstack [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.000894] env[65385]: DEBUG nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1095.049136] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454411, 'name': CreateVM_Task, 'duration_secs': 0.615243} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.052295] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1095.054579] env[65385]: DEBUG nova.network.neutron [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updated VIF entry in instance network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1095.054952] env[65385]: DEBUG nova.network.neutron [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1095.056723] env[65385]: WARNING neutronclient.v2_0.client [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.057090] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.057236] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.057546] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1095.058317] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1af545b-b892-460d-ad23-760491b83b70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.064250] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1095.064250] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5271b5bd-09f8-6386-d221-9252bb180dab" [ 1095.064250] env[65385]: _type = "Task" [ 1095.064250] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.076694] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5271b5bd-09f8-6386-d221-9252bb180dab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.086827] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1095.087300] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6c44282-89eb-448e-827e-93cd65e46abc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.095349] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1095.095349] env[65385]: value = "task-4454412" [ 1095.095349] env[65385]: _type = "Task" [ 1095.095349] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.104307] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454412, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.118581] env[65385]: WARNING neutronclient.v2_0.client [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.119414] env[65385]: WARNING openstack [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.119851] env[65385]: WARNING openstack [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.219921] env[65385]: DEBUG nova.network.neutron [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [{"id": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "address": "fa:16:3e:49:27:54", "network": {"id": "380eb239-46a9-4bb2-be3e-66bb113acf76", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1139908934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0a282a49e5f44dc781471d9c2103ca8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38f5da-81", "ovs_interfaceid": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1095.284049] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277524d3-3dc9-4021-8a87-18b77686079f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.296908] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b913a3cd-b24b-4d5c-befe-ece4785389aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.330277] env[65385]: DEBUG nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1095.333564] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ae5d3a-864f-4711-9c2f-aa1611c1e251 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.343219] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fa3a37-8b69-419c-bd8f-a2c5075d2f1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.358915] env[65385]: DEBUG nova.compute.provider_tree [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.560044] env[65385]: DEBUG oslo_concurrency.lockutils [req-cc3b52b0-dc1b-45d0-bd7d-fdae7ea63e79 req-636f6a2c-c4e3-462c-9311-29590e6f3385 service nova] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.576157] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5271b5bd-09f8-6386-d221-9252bb180dab, 'name': SearchDatastore_Task, 'duration_secs': 0.010763} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.576529] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.576804] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.577140] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.577322] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.577521] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.577834] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-978b9cda-2c53-434e-9c3f-8a0d28f384f3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.589336] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.589548] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1095.590560] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad05e6d4-0d3f-4678-8399-01571a34b750 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.597936] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1095.597936] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52461f99-bcaa-c083-4efb-1b3004ec5645" [ 1095.597936] env[65385]: _type = "Task" [ 1095.597936] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.612654] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52461f99-bcaa-c083-4efb-1b3004ec5645, 'name': SearchDatastore_Task, 'duration_secs': 0.011999} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.616940] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454412, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.617237] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-787a95cd-f577-4c82-aa0a-963370e6eb07 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.624506] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1095.624506] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dd5dd9-e5c5-4ea5-02c5-1c0e74f74554" [ 1095.624506] env[65385]: _type = "Task" [ 1095.624506] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.638188] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dd5dd9-e5c5-4ea5-02c5-1c0e74f74554, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.722680] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Releasing lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.858591] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.861798] env[65385]: DEBUG nova.scheduler.client.report [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.016802] env[65385]: DEBUG nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1096.039862] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1096.040135] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1096.040280] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1096.040462] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1096.040598] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1096.040739] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1096.040945] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1096.041118] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1096.041277] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1096.041435] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1096.041691] env[65385]: DEBUG nova.virt.hardware [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1096.042655] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b41136-fc17-4204-b6e0-e1fc3dc5c905 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.052742] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac32438-de58-44ec-b3fb-7b84be123c1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.107201] env[65385]: DEBUG oslo_vmware.api [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454412, 'name': PowerOnVM_Task, 'duration_secs': 0.742364} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.107467] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1096.107611] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf8d990-b1fa-4e86-ab54-315aac1c9b18 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance '29a449d0-f62d-4430-96ab-9ac43e5a7ad2' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1096.135767] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dd5dd9-e5c5-4ea5-02c5-1c0e74f74554, 'name': SearchDatastore_Task, 'duration_secs': 0.011986} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.136211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.136806] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f/b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1096.136952] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ae230dc-9a3b-441a-81e9-70ea4e942c1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.146248] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1096.146248] env[65385]: value = "task-4454413" [ 1096.146248] env[65385]: _type = "Task" [ 1096.146248] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.156264] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.366809] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.367489] env[65385]: DEBUG nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1096.370473] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.886s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.402351] env[65385]: DEBUG nova.compute.manager [req-6b732487-2a13-498c-8e71-a74834808013 req-a0e9af2b-1398-4b06-b4a3-ca8c95cdf6c2 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-vif-plugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1096.402698] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b732487-2a13-498c-8e71-a74834808013 req-a0e9af2b-1398-4b06-b4a3-ca8c95cdf6c2 service nova] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.403461] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b732487-2a13-498c-8e71-a74834808013 req-a0e9af2b-1398-4b06-b4a3-ca8c95cdf6c2 service nova] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.403461] env[65385]: DEBUG oslo_concurrency.lockutils [req-6b732487-2a13-498c-8e71-a74834808013 req-a0e9af2b-1398-4b06-b4a3-ca8c95cdf6c2 service nova] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.403583] env[65385]: DEBUG nova.compute.manager [req-6b732487-2a13-498c-8e71-a74834808013 req-a0e9af2b-1398-4b06-b4a3-ca8c95cdf6c2 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] No waiting events found dispatching network-vif-plugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1096.403871] env[65385]: WARNING nova.compute.manager [req-6b732487-2a13-498c-8e71-a74834808013 req-a0e9af2b-1398-4b06-b4a3-ca8c95cdf6c2 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received unexpected event network-vif-plugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 for instance with vm_state building and task_state spawning. [ 1096.510531] env[65385]: DEBUG nova.network.neutron [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Successfully updated port: cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1096.657915] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454413, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493986} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.658444] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f/b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1096.658668] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.658929] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c886c1b8-e1aa-4385-a37a-b8112c832141 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.667699] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1096.667699] env[65385]: value = "task-4454414" [ 1096.667699] env[65385]: _type = "Task" [ 1096.667699] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.681323] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.873812] env[65385]: DEBUG nova.compute.utils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1096.877416] env[65385]: INFO nova.compute.claims [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1096.881164] env[65385]: DEBUG nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1096.881436] env[65385]: DEBUG nova.network.neutron [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1096.881721] env[65385]: WARNING neutronclient.v2_0.client [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.882117] env[65385]: WARNING neutronclient.v2_0.client [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.882766] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.883085] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.942298] env[65385]: DEBUG nova.policy [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662871f375e441eb98d814cf7d7ac5c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bc8afc699e34f059f18d8244980bc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1097.013638] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.013881] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.014112] env[65385]: DEBUG nova.network.neutron [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1097.179016] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071458} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.179359] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1097.180233] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe80369-afb9-4e3c-9c6d-18aa292e52a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.204844] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f/b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1097.205274] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bed9206f-83ce-4bb8-b024-842a8f674b3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.229026] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1097.229026] env[65385]: value = "task-4454415" [ 1097.229026] env[65385]: _type = "Task" [ 1097.229026] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.237487] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454415, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.245985] env[65385]: DEBUG nova.network.neutron [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Successfully created port: e4c07525-65a3-431c-a14b-f434713c739f {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1097.263661] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.264044] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09aabe8a-763d-45ae-9716-82a4a024c2e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.271979] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1097.271979] env[65385]: value = "task-4454416" [ 1097.271979] env[65385]: _type = "Task" [ 1097.271979] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.280868] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454416, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.393386] env[65385]: INFO nova.compute.resource_tracker [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating resource usage from migration 3e37f651-5be2-4236-9a1c-b5bd5eb2e876 [ 1097.401193] env[65385]: DEBUG nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1097.516929] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.517315] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.558824] env[65385]: DEBUG nova.network.neutron [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1097.579697] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.580035] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.646512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235a6081-fa2c-4bc3-b2d8-2061c6bb4913 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.654914] env[65385]: WARNING neutronclient.v2_0.client [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.655583] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.655929] env[65385]: WARNING openstack [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.664087] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0627e306-db3c-4a31-bab4-a85f80d82eaa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.707424] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6fc87c-ff63-417d-a99c-7f1ff7868bfd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.720582] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b50eef8-1b30-4d59-bc17-779d010c69e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.739012] env[65385]: DEBUG nova.compute.provider_tree [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.746572] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454415, 'name': ReconfigVM_Task, 'duration_secs': 0.386556} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.746572] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Reconfigured VM instance instance-00000069 to attach disk [datastore1] b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f/b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.747737] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf32b5bf-398a-4b39-ac2e-5695075c376a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.753074] env[65385]: DEBUG nova.network.neutron [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd731def-a6", "ovs_interfaceid": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.760265] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1097.760265] env[65385]: value = "task-4454417" [ 1097.760265] env[65385]: _type = "Task" [ 1097.760265] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.769638] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454417, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.781619] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454416, 'name': PowerOffVM_Task, 'duration_secs': 0.197533} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.781933] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.782853] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54412bc-1d05-4a67-9865-e2fb53be0067 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.803792] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b37561-eec1-4426-88cc-93d316bdb51d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.846337] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.846655] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-945c6541-8ea6-428e-95be-7b3b8916dc5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.857054] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1097.857054] env[65385]: value = "task-4454418" [ 1097.857054] env[65385]: _type = "Task" [ 1097.857054] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.866881] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1097.867227] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.867505] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.867648] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.867835] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.868110] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75cca18b-1f2b-43a7-bd1d-468932a53503 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.878651] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.878865] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1097.879606] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-968d2619-7d22-414c-aad2-7a2d511d89b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.889061] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1097.889061] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c86b86-70ee-aeb3-039e-2062951967ad" [ 1097.889061] env[65385]: _type = "Task" [ 1097.889061] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.899368] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c86b86-70ee-aeb3-039e-2062951967ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.241983] env[65385]: DEBUG nova.scheduler.client.report [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1098.254635] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.254890] env[65385]: DEBUG nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Instance network_info: |[{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd731def-a6", "ovs_interfaceid": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1098.255540] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:32:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4a8fd90-153b-494f-b76a-299eb05c03f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd731def-a6f7-40f9-b6ad-a10d2fd72984', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1098.264088] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1098.265140] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1098.268600] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-983717f6-a4e1-43e6-94b8-2c0db38d9767 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.291728] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454417, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.293530] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1098.293530] env[65385]: value = "task-4454419" [ 1098.293530] env[65385]: _type = "Task" [ 1098.293530] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.303765] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454419, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.402959] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c86b86-70ee-aeb3-039e-2062951967ad, 'name': SearchDatastore_Task, 'duration_secs': 0.01865} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.403899] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dea09513-3447-4bcc-8e49-be85aa3668fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.410484] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1098.410484] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e9865-91b3-3322-9cfb-582cd3a4b2ea" [ 1098.410484] env[65385]: _type = "Task" [ 1098.410484] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.411686] env[65385]: DEBUG nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1098.423966] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e9865-91b3-3322-9cfb-582cd3a4b2ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.438350] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1098.439037] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1098.439037] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1098.439037] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1098.439167] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1098.439315] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1098.439609] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1098.439795] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1098.439971] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1098.440159] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1098.440373] env[65385]: DEBUG nova.virt.hardware [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1098.441393] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873b0de9-d08d-452b-a32e-41e7aa63df41 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.449010] env[65385]: DEBUG nova.compute.manager [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1098.449340] env[65385]: DEBUG nova.compute.manager [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing instance network info cache due to event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1098.449340] env[65385]: DEBUG oslo_concurrency.lockutils [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Acquiring lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.449542] env[65385]: DEBUG oslo_concurrency.lockutils [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Acquired lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.450532] env[65385]: DEBUG nova.network.neutron [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1098.458450] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9b2d07-7fb7-4522-a909-82e9e35ac00c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.467724] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.590873] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.591594] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.592167] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.669473] env[65385]: DEBUG nova.network.neutron [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Port ace973a7-f2da-4905-99cb-e72ce9ef3f18 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1098.669473] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.669630] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.669664] env[65385]: DEBUG nova.network.neutron [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1098.738552] env[65385]: DEBUG nova.compute.manager [req-53bdc4a0-b5ca-4aa1-9e7a-dfdc6fa8bbf4 req-16bf8888-beac-4fa5-971a-225e3b5596c2 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Received event network-vif-plugged-e4c07525-65a3-431c-a14b-f434713c739f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1098.738763] env[65385]: DEBUG oslo_concurrency.lockutils [req-53bdc4a0-b5ca-4aa1-9e7a-dfdc6fa8bbf4 req-16bf8888-beac-4fa5-971a-225e3b5596c2 service nova] Acquiring lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.738962] env[65385]: DEBUG oslo_concurrency.lockutils [req-53bdc4a0-b5ca-4aa1-9e7a-dfdc6fa8bbf4 req-16bf8888-beac-4fa5-971a-225e3b5596c2 service nova] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.739140] env[65385]: DEBUG oslo_concurrency.lockutils [req-53bdc4a0-b5ca-4aa1-9e7a-dfdc6fa8bbf4 req-16bf8888-beac-4fa5-971a-225e3b5596c2 service nova] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.739301] env[65385]: DEBUG nova.compute.manager [req-53bdc4a0-b5ca-4aa1-9e7a-dfdc6fa8bbf4 req-16bf8888-beac-4fa5-971a-225e3b5596c2 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] No waiting events found dispatching network-vif-plugged-e4c07525-65a3-431c-a14b-f434713c739f {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1098.739461] env[65385]: WARNING nova.compute.manager [req-53bdc4a0-b5ca-4aa1-9e7a-dfdc6fa8bbf4 req-16bf8888-beac-4fa5-971a-225e3b5596c2 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Received unexpected event network-vif-plugged-e4c07525-65a3-431c-a14b-f434713c739f for instance with vm_state building and task_state spawning. [ 1098.747212] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.377s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.747432] env[65385]: INFO nova.compute.manager [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Migrating [ 1098.755015] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.050s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.755275] env[65385]: DEBUG nova.objects.instance [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1098.776875] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454417, 'name': Rename_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.807722] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454419, 'name': CreateVM_Task, 'duration_secs': 0.379573} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.807920] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1098.808651] env[65385]: WARNING neutronclient.v2_0.client [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.809211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.809434] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.809892] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1098.810260] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66152d05-a991-48dd-80c7-f16eb0454d79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.815283] env[65385]: DEBUG nova.network.neutron [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Successfully updated port: e4c07525-65a3-431c-a14b-f434713c739f {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1098.818044] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1098.818044] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52141b48-9c52-5d5d-d7e0-e7ec03ee1f26" [ 1098.818044] env[65385]: _type = "Task" [ 1098.818044] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.829710] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52141b48-9c52-5d5d-d7e0-e7ec03ee1f26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.921406] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525e9865-91b3-3322-9cfb-582cd3a4b2ea, 'name': SearchDatastore_Task, 'duration_secs': 0.01325} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.921874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.922206] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. {{(pid=65385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1098.922504] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff3f2066-6a8e-45f4-9071-94a260f4afb3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.930387] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1098.930387] env[65385]: value = "task-4454420" [ 1098.930387] env[65385]: _type = "Task" [ 1098.930387] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.939443] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454420, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.953613] env[65385]: WARNING neutronclient.v2_0.client [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.954477] env[65385]: WARNING openstack [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.954932] env[65385]: WARNING openstack [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.089536] env[65385]: WARNING openstack [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.089971] env[65385]: WARNING openstack [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.160504] env[65385]: WARNING neutronclient.v2_0.client [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.161250] env[65385]: WARNING openstack [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.161581] env[65385]: WARNING openstack [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.172943] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.173651] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.174055] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.260989] env[65385]: DEBUG nova.network.neutron [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updated VIF entry in instance network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1099.261482] env[65385]: DEBUG nova.network.neutron [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd731def-a6", "ovs_interfaceid": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1099.267444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.267444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.267636] env[65385]: DEBUG nova.network.neutron [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1099.281040] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454417, 'name': Rename_Task, 'duration_secs': 1.201823} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.281382] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.281719] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9531b09c-4a0d-4c69-b682-899bcc1e1ca8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.292983] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1099.292983] env[65385]: value = "task-4454421" [ 1099.292983] env[65385]: _type = "Task" [ 1099.292983] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.300628] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.301299] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.314288] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454421, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.319235] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "refresh_cache-7f22736b-654f-4d51-82d4-3fe80c1d0b52" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.319491] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "refresh_cache-7f22736b-654f-4d51-82d4-3fe80c1d0b52" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.319806] env[65385]: DEBUG nova.network.neutron [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1099.333999] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52141b48-9c52-5d5d-d7e0-e7ec03ee1f26, 'name': SearchDatastore_Task, 'duration_secs': 0.01202} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.335065] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.335376] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1099.335616] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.335807] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.336051] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1099.336659] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-100e89b9-8ce9-4081-a3f3-bd5595ad43d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.360899] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1099.361134] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1099.362156] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad741da4-7012-49fb-8a25-37e46ad02aa9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.372178] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1099.372178] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52436b5e-0613-6892-d28d-ab5b685be216" [ 1099.372178] env[65385]: _type = "Task" [ 1099.372178] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.379532] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.380208] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.380554] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.392296] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52436b5e-0613-6892-d28d-ab5b685be216, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.442081] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454420, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.493935] env[65385]: DEBUG nova.network.neutron [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1099.767699] env[65385]: DEBUG oslo_concurrency.lockutils [req-c265c3b5-576e-412e-93bd-9040d9d3a829 req-2c7bb160-1288-4675-952f-d074583ed234 service nova] Releasing lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.770914] env[65385]: DEBUG oslo_concurrency.lockutils [None req-05b96971-3ab5-4b12-a81d-cae666b53290 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.772205] env[65385]: WARNING neutronclient.v2_0.client [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.772850] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.773199] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.780977] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.234s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.781755] env[65385]: DEBUG nova.objects.instance [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'resources' on Instance uuid 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.803973] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454421, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.826891] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.827327] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.863875] env[65385]: DEBUG nova.network.neutron [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1099.879561] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.880052] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.891596] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.891992] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.898554] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52436b5e-0613-6892-d28d-ab5b685be216, 'name': SearchDatastore_Task, 'duration_secs': 0.050587} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.899426] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eed47f83-61e8-4aff-ba51-ad690d7f2a4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.906236] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1099.906236] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279b9b1-65d9-e75d-3f19-8232e736916d" [ 1099.906236] env[65385]: _type = "Task" [ 1099.906236] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.915430] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279b9b1-65d9-e75d-3f19-8232e736916d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.941874] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454420, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534158} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.942626] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. [ 1099.942872] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4a9a32-8e32-485b-92b1-2ad256af4426 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.968482] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.969787] env[65385]: WARNING neutronclient.v2_0.client [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.970421] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.970766] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.978819] env[65385]: WARNING neutronclient.v2_0.client [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.979367] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.979690] env[65385]: WARNING openstack [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.986525] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-684a126c-f230-4c15-952e-dfd8e56c33e2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.999924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.007197] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1100.007197] env[65385]: value = "task-4454422" [ 1100.007197] env[65385]: _type = "Task" [ 1100.007197] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.018783] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454422, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.088709] env[65385]: DEBUG nova.network.neutron [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance_info_cache with network_info: [{"id": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "address": "fa:16:3e:47:9a:0a", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped7c17c7-54", "ovs_interfaceid": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1100.092128] env[65385]: DEBUG nova.network.neutron [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Updating instance_info_cache with network_info: [{"id": "e4c07525-65a3-431c-a14b-f434713c739f", "address": "fa:16:3e:cd:5d:a2", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c07525-65", "ovs_interfaceid": "e4c07525-65a3-431c-a14b-f434713c739f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1100.304929] env[65385]: DEBUG oslo_vmware.api [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454421, 'name': PowerOnVM_Task, 'duration_secs': 0.77841} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.307709] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1100.307926] env[65385]: INFO nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Took 8.51 seconds to spawn the instance on the hypervisor. [ 1100.308117] env[65385]: DEBUG nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1100.309279] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ea9d8a-3a40-46a7-a5ad-8adb9b7939b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.418490] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5279b9b1-65d9-e75d-3f19-8232e736916d, 'name': SearchDatastore_Task, 'duration_secs': 0.033557} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.418757] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.419025] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1100.419299] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7c5d74a-2782-48eb-b005-5aae482120dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.428071] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1100.428071] env[65385]: value = "task-4454423" [ 1100.428071] env[65385]: _type = "Task" [ 1100.428071] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.442465] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454423, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.506363] env[65385]: DEBUG nova.compute.manager [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65385) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:925}} [ 1100.509835] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04235734-13f1-45bc-abf2-365bc6475f61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.526431] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.529453] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f579cc-7112-4754-a514-f4d3a46da342 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.564960] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9310fb4c-b061-4f7b-bacd-3fb7fc554cb4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.576238] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e735c79-6cd0-4759-9243-1fe739581f34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.593365] env[65385]: DEBUG nova.compute.provider_tree [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.597758] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.599043] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "refresh_cache-7f22736b-654f-4d51-82d4-3fe80c1d0b52" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.599352] env[65385]: DEBUG nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Instance network_info: |[{"id": "e4c07525-65a3-431c-a14b-f434713c739f", "address": "fa:16:3e:cd:5d:a2", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c07525-65", "ovs_interfaceid": "e4c07525-65a3-431c-a14b-f434713c739f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1100.601191] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:5d:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4c07525-65a3-431c-a14b-f434713c739f', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1100.608189] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1100.609008] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1100.609250] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c013798-1061-4d2b-a5bb-c2b0bedb4879 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.635817] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1100.635817] env[65385]: value = "task-4454424" [ 1100.635817] env[65385]: _type = "Task" [ 1100.635817] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.646566] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454424, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.835466] env[65385]: INFO nova.compute.manager [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Took 14.70 seconds to build instance. [ 1100.943276] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454423, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.020544] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454422, 'name': ReconfigVM_Task, 'duration_secs': 0.694268} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.020781] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Reconfigured VM instance instance-00000068 to attach disk [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1101.021739] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e04dc3-d813-449c-b482-e06d3f7234c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.049331] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c6d2589-eb4f-477d-aadc-fccff3c5231b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.065534] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1101.065534] env[65385]: value = "task-4454425" [ 1101.065534] env[65385]: _type = "Task" [ 1101.065534] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.074272] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454425, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.104074] env[65385]: DEBUG nova.scheduler.client.report [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1101.147899] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454424, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.155081] env[65385]: DEBUG nova.compute.manager [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Received event network-changed-e4c07525-65a3-431c-a14b-f434713c739f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1101.155173] env[65385]: DEBUG nova.compute.manager [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Refreshing instance network info cache due to event network-changed-e4c07525-65a3-431c-a14b-f434713c739f. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1101.155370] env[65385]: DEBUG oslo_concurrency.lockutils [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Acquiring lock "refresh_cache-7f22736b-654f-4d51-82d4-3fe80c1d0b52" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.155488] env[65385]: DEBUG oslo_concurrency.lockutils [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Acquired lock "refresh_cache-7f22736b-654f-4d51-82d4-3fe80c1d0b52" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.155642] env[65385]: DEBUG nova.network.neutron [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Refreshing network info cache for port e4c07525-65a3-431c-a14b-f434713c739f {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1101.335776] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7539690e-f095-4cd5-ac61-434a2fe8b107 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.210s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.443791] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454423, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592676} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.443791] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1101.443791] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1101.444228] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f8a174e-7933-4a8f-9d3d-c7af42d56165 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.452258] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1101.452258] env[65385]: value = "task-4454426" [ 1101.452258] env[65385]: _type = "Task" [ 1101.452258] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.462156] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454426, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.578110] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454425, 'name': ReconfigVM_Task, 'duration_secs': 0.354573} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.578404] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1101.578671] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae01b322-eddc-4556-81a2-19e612a0eca8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.587589] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1101.587589] env[65385]: value = "task-4454427" [ 1101.587589] env[65385]: _type = "Task" [ 1101.587589] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.598417] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454427, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.612479] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.617997] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.759s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.619519] env[65385]: INFO nova.compute.claims [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1101.649266] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454424, 'name': CreateVM_Task, 'duration_secs': 0.633589} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.649673] env[65385]: INFO nova.scheduler.client.report [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted allocations for instance 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c [ 1101.650706] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1101.651488] env[65385]: WARNING neutronclient.v2_0.client [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.652358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.652358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.652358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1101.654939] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74d9a32a-5e17-4ef8-8108-b8201b81fbe5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.659107] env[65385]: WARNING neutronclient.v2_0.client [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.660025] env[65385]: WARNING openstack [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.660025] env[65385]: WARNING openstack [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.670694] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1101.670694] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52efa222-942a-77cd-22e1-1d2928ec595a" [ 1101.670694] env[65385]: _type = "Task" [ 1101.670694] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.676074] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.683764] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52efa222-942a-77cd-22e1-1d2928ec595a, 'name': SearchDatastore_Task, 'duration_secs': 0.012099} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.684102] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.684344] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1101.684575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.684730] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.684886] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1101.685182] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54609f3b-84c0-44fe-a36e-3f0a7817415e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.695497] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1101.695684] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1101.696589] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2447a9f-f01c-4d66-9d59-2af0ea3d3e15 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.705031] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1101.705031] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ec9bd9-b78d-390f-518b-b90523af5f43" [ 1101.705031] env[65385]: _type = "Task" [ 1101.705031] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.714248] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ec9bd9-b78d-390f-518b-b90523af5f43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.804356] env[65385]: WARNING openstack [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.804592] env[65385]: WARNING openstack [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.887565] env[65385]: WARNING neutronclient.v2_0.client [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.889657] env[65385]: WARNING openstack [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.889953] env[65385]: WARNING openstack [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.963204] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454426, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064433} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.963495] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.964400] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e9bb32-0869-4ae4-b6d8-9c90052eb5c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.987870] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.988687] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-889f583a-4c6e-4dcd-af02-e0709413fb15 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.005173] env[65385]: DEBUG nova.network.neutron [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Updated VIF entry in instance network info cache for port e4c07525-65a3-431c-a14b-f434713c739f. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1102.005528] env[65385]: DEBUG nova.network.neutron [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Updating instance_info_cache with network_info: [{"id": "e4c07525-65a3-431c-a14b-f434713c739f", "address": "fa:16:3e:cd:5d:a2", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4c07525-65", "ovs_interfaceid": "e4c07525-65a3-431c-a14b-f434713c739f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1102.012707] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1102.012707] env[65385]: value = "task-4454428" [ 1102.012707] env[65385]: _type = "Task" [ 1102.012707] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.022904] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454428, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.100855] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454427, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.129440] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df14fc2e-0e62-4311-ade8-3dfa546bfbdd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.152563] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance '805528b8-96db-4bb8-ba05-2b6332c54624' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1102.162032] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d9be60a5-31f1-4ae4-baa3-12dd184c39fd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "5ae074b9-4e82-4345-9bc1-ab8d7c930f1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.514s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.216125] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ec9bd9-b78d-390f-518b-b90523af5f43, 'name': SearchDatastore_Task, 'duration_secs': 0.011768} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.216971] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d604475-e9d4-4375-a0b2-ead4c9397242 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.223041] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1102.223041] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa65a0-6dcc-ae68-1940-0c8715eff63b" [ 1102.223041] env[65385]: _type = "Task" [ 1102.223041] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.231550] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa65a0-6dcc-ae68-1940-0c8715eff63b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.508464] env[65385]: DEBUG oslo_concurrency.lockutils [req-2263677d-c1ef-440f-afce-bf1120a19d4a req-f8a888e0-1fbe-43a6-b0d1-96152f37bd12 service nova] Releasing lock "refresh_cache-7f22736b-654f-4d51-82d4-3fe80c1d0b52" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.525087] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454428, 'name': ReconfigVM_Task, 'duration_secs': 0.475608} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.525424] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Reconfigured VM instance instance-0000006a to attach disk [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.526205] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6419d10-f816-43dd-90fa-382cc7ca0d0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.533295] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1102.533295] env[65385]: value = "task-4454429" [ 1102.533295] env[65385]: _type = "Task" [ 1102.533295] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.544333] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454429, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.601570] env[65385]: DEBUG oslo_vmware.api [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454427, 'name': PowerOnVM_Task, 'duration_secs': 0.937244} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.602511] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1102.606364] env[65385]: DEBUG nova.compute.manager [None req-00817d8d-2c9d-497e-967e-7888c21678a4 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1102.607338] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c110b56d-dafc-4b3b-816a-1609f54f5193 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.665381] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.665381] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ace02c91-b19e-441b-a71f-f3305d40a6b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.677031] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1102.677031] env[65385]: value = "task-4454430" [ 1102.677031] env[65385]: _type = "Task" [ 1102.677031] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.696287] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.739867] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52aa65a0-6dcc-ae68-1940-0c8715eff63b, 'name': SearchDatastore_Task, 'duration_secs': 0.014315} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.740236] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.741124] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 7f22736b-654f-4d51-82d4-3fe80c1d0b52/7f22736b-654f-4d51-82d4-3fe80c1d0b52.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1102.741124] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fc47d15-5215-4d84-b9f3-e1c21532cefa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.753791] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1102.753791] env[65385]: value = "task-4454431" [ 1102.753791] env[65385]: _type = "Task" [ 1102.753791] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.764231] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.925119] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab24254-e090-481c-a116-d629c33a2eb0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.936355] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5294f6b8-3284-457d-80f2-1fc9d2e02706 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.973917] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dba01a-2d7b-428d-a628-2016c2a073cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.983574] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d84b42-adab-45d3-b1e5-6f79ab9f4e56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.002599] env[65385]: DEBUG nova.compute.provider_tree [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.046272] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454429, 'name': Rename_Task, 'duration_secs': 0.173622} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.046614] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1103.046980] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a0e37ca-ab09-4347-985b-3374363bae10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.056879] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1103.056879] env[65385]: value = "task-4454432" [ 1103.056879] env[65385]: _type = "Task" [ 1103.056879] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.067107] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.189676] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454430, 'name': PowerOffVM_Task, 'duration_secs': 0.28063} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.190312] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.190791] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance '805528b8-96db-4bb8-ba05-2b6332c54624' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1103.268421] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454431, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.305755] env[65385]: DEBUG nova.compute.manager [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1103.306031] env[65385]: DEBUG nova.compute.manager [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing instance network info cache due to event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1103.306360] env[65385]: DEBUG oslo_concurrency.lockutils [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.306401] env[65385]: DEBUG oslo_concurrency.lockutils [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.307307] env[65385]: DEBUG nova.network.neutron [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1103.506094] env[65385]: DEBUG nova.scheduler.client.report [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1103.569520] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454432, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.698808] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1103.699070] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1103.699228] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1103.699461] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1103.699610] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1103.699752] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1103.699955] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.700126] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1103.700291] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1103.700447] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1103.700695] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1103.706336] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e005e1f-88cf-481a-ae6c-7c92dbf780f0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.727817] env[65385]: DEBUG nova.compute.manager [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1103.728130] env[65385]: DEBUG nova.compute.manager [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing instance network info cache due to event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1103.728383] env[65385]: DEBUG oslo_concurrency.lockutils [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.728569] env[65385]: DEBUG oslo_concurrency.lockutils [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.728763] env[65385]: DEBUG nova.network.neutron [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1103.732206] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1103.732206] env[65385]: value = "task-4454433" [ 1103.732206] env[65385]: _type = "Task" [ 1103.732206] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.744387] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454433, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.771446] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561291} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.772364] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 7f22736b-654f-4d51-82d4-3fe80c1d0b52/7f22736b-654f-4d51-82d4-3fe80c1d0b52.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1103.772364] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1103.772364] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6cd69734-47bf-40ed-8d31-28abbc6a1bdc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.780607] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1103.780607] env[65385]: value = "task-4454434" [ 1103.780607] env[65385]: _type = "Task" [ 1103.780607] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.792949] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454434, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.810292] env[65385]: WARNING neutronclient.v2_0.client [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.811391] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.811882] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.011581] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.012104] env[65385]: DEBUG nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1104.015444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.340s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.072176] env[65385]: DEBUG oslo_vmware.api [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454432, 'name': PowerOnVM_Task, 'duration_secs': 0.756204} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.072517] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.072815] env[65385]: INFO nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Took 8.06 seconds to spawn the instance on the hypervisor. [ 1104.073025] env[65385]: DEBUG nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1104.074467] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cf33f1-16e0-4067-8c18-4f6bc2283eff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.128911] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.128911] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.225109] env[65385]: WARNING neutronclient.v2_0.client [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.225772] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.226174] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.235644] env[65385]: WARNING neutronclient.v2_0.client [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.235752] env[65385]: WARNING openstack [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.236107] env[65385]: WARNING openstack [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.255331] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454433, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.292012] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454434, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.232289} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.292340] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1104.293351] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3934792b-5b1a-4eb1-b51a-49e1606fa5de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.317258] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 7f22736b-654f-4d51-82d4-3fe80c1d0b52/7f22736b-654f-4d51-82d4-3fe80c1d0b52.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.326593] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f65e1c99-c3e5-499d-b6d6-73e46d20edff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.347501] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1104.347501] env[65385]: value = "task-4454435" [ 1104.347501] env[65385]: _type = "Task" [ 1104.347501] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.357299] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454435, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.365334] env[65385]: WARNING openstack [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.365817] env[65385]: WARNING openstack [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.402636] env[65385]: DEBUG nova.network.neutron [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updated VIF entry in instance network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1104.406152] env[65385]: DEBUG nova.network.neutron [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.456505] env[65385]: WARNING neutronclient.v2_0.client [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.457430] env[65385]: WARNING openstack [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.457783] env[65385]: WARNING openstack [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.526028] env[65385]: DEBUG nova.compute.utils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1104.526787] env[65385]: DEBUG nova.objects.instance [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'migration_context' on Instance uuid 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.531081] env[65385]: DEBUG nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1104.531704] env[65385]: DEBUG nova.network.neutron [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1104.532437] env[65385]: WARNING neutronclient.v2_0.client [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.533494] env[65385]: WARNING neutronclient.v2_0.client [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.535209] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.535710] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.566129] env[65385]: DEBUG nova.network.neutron [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updated VIF entry in instance network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1104.566717] env[65385]: DEBUG nova.network.neutron [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.602325] env[65385]: INFO nova.compute.manager [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Took 14.68 seconds to build instance. [ 1104.651501] env[65385]: DEBUG nova.policy [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03f7c78b16e547c38f24ed2c8a329202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ba61d77cc2d4fe1ba3c03466d5985d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1104.758165] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454433, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.859840] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454435, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.907965] env[65385]: DEBUG oslo_concurrency.lockutils [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.907965] env[65385]: DEBUG nova.compute.manager [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1104.907965] env[65385]: DEBUG nova.compute.manager [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing instance network info cache due to event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1104.907965] env[65385]: DEBUG oslo_concurrency.lockutils [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.027992] env[65385]: DEBUG nova.network.neutron [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Successfully created port: 2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1105.053220] env[65385]: DEBUG nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1105.071331] env[65385]: DEBUG oslo_concurrency.lockutils [req-afb8ef36-3f37-4b11-825d-5ed758fa31dc req-d386ca3d-b6d4-4232-9b3a-1114af61084f service nova] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1105.071331] env[65385]: DEBUG oslo_concurrency.lockutils [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1105.071331] env[65385]: DEBUG nova.network.neutron [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1105.105853] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89f92852-ecfe-42a2-b32a-806213a23371 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.188s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.168016] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.168265] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.258432] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454433, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.313027] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d5c21b-18d1-438d-bfa9-c97e50fe5c4a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.325027] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf135257-360e-4df1-877c-5b102cd1a6eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.361788] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a6984c-99a6-4b79-acb2-19fbf1a726f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.375358] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42115e9-f716-4171-867c-c836b9c4bd30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.381730] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454435, 'name': ReconfigVM_Task, 'duration_secs': 0.534075} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.382109] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 7f22736b-654f-4d51-82d4-3fe80c1d0b52/7f22736b-654f-4d51-82d4-3fe80c1d0b52.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.383183] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02b49fcb-223c-4dde-b320-e8340fad2eca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.404790] env[65385]: DEBUG nova.compute.provider_tree [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1105.406132] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1105.406132] env[65385]: value = "task-4454436" [ 1105.406132] env[65385]: _type = "Task" [ 1105.406132] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.416283] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454436, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.575688] env[65385]: WARNING neutronclient.v2_0.client [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.576530] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.576916] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.676146] env[65385]: DEBUG nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1105.756985] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454433, 'name': ReconfigVM_Task, 'duration_secs': 1.58797} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.757335] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance '805528b8-96db-4bb8-ba05-2b6332c54624' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1105.908488] env[65385]: DEBUG nova.scheduler.client.report [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1105.922660] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454436, 'name': Rename_Task, 'duration_secs': 0.26163} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.923261] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1105.923261] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67eb6147-7251-4441-b0d1-7398c7879ea1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.931514] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1105.931514] env[65385]: value = "task-4454437" [ 1105.931514] env[65385]: _type = "Task" [ 1105.931514] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.941809] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.064472] env[65385]: DEBUG nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1106.122621] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1106.122621] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1106.122792] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1106.122885] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1106.123260] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1106.123860] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1106.123860] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.126535] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1106.126535] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1106.126535] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1106.126535] env[65385]: DEBUG nova.virt.hardware [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1106.126535] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452b8f16-bf4b-4d6b-9bde-8ae9bc8e3cdf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.136397] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915a66bb-abe9-40b7-9ea0-ff55413b6159 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.209031] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.213842] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1106.214240] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1106.265431] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1106.265431] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1106.265687] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1106.265928] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1106.266191] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1106.266357] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1106.266572] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.266729] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1106.266903] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1106.267062] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1106.267454] env[65385]: DEBUG nova.virt.hardware [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1106.274387] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1106.274796] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fb61893-16bb-486b-9dd0-f8b56a4e72d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.296823] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1106.296823] env[65385]: value = "task-4454438" [ 1106.296823] env[65385]: _type = "Task" [ 1106.296823] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.309725] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.444327] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454437, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.637968] env[65385]: DEBUG nova.network.neutron [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Successfully updated port: 2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1106.674714] env[65385]: WARNING neutronclient.v2_0.client [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1106.675592] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1106.676273] env[65385]: WARNING openstack [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1106.689336] env[65385]: INFO nova.compute.manager [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Rebuilding instance [ 1106.780809] env[65385]: DEBUG nova.compute.manager [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1106.782207] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd885d91-0a54-4122-bd98-64001f6757c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.807522] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454438, 'name': ReconfigVM_Task, 'duration_secs': 0.193823} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.807522] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1106.808207] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714f4dd3-f50d-42f4-a4e4-40741b24d4a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.832282] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 805528b8-96db-4bb8-ba05-2b6332c54624/805528b8-96db-4bb8-ba05-2b6332c54624.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.832625] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85ff14ad-adde-40cc-b652-39bac8bd15bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.849098] env[65385]: DEBUG nova.network.neutron [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updated VIF entry in instance network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1106.849614] env[65385]: DEBUG nova.network.neutron [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1106.858181] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1106.858181] env[65385]: value = "task-4454439" [ 1106.858181] env[65385]: _type = "Task" [ 1106.858181] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.868077] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454439, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.901413] env[65385]: DEBUG nova.compute.manager [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1106.901651] env[65385]: DEBUG nova.compute.manager [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing instance network info cache due to event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1106.901909] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.902085] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.902885] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1106.920371] env[65385]: DEBUG nova.compute.manager [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1106.920539] env[65385]: DEBUG nova.compute.manager [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing instance network info cache due to event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1106.920750] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Acquiring lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.920964] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Acquired lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.921159] env[65385]: DEBUG nova.network.neutron [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1106.924410] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.909s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1106.932196] env[65385]: WARNING neutronclient.v2_0.client [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1106.932732] env[65385]: WARNING openstack [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1106.933164] env[65385]: WARNING openstack [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1106.940331] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.731s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1106.941972] env[65385]: INFO nova.compute.claims [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1106.962810] env[65385]: DEBUG oslo_vmware.api [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454437, 'name': PowerOnVM_Task, 'duration_secs': 0.679056} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.962965] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.963283] env[65385]: INFO nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Took 8.55 seconds to spawn the instance on the hypervisor. [ 1106.963458] env[65385]: DEBUG nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1106.964394] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e02c684-6f34-4038-9240-b4b5301bd11b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.082186] env[65385]: WARNING openstack [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.082871] env[65385]: WARNING openstack [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.144453] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.144453] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1107.144453] env[65385]: DEBUG nova.network.neutron [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1107.149141] env[65385]: WARNING neutronclient.v2_0.client [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.149843] env[65385]: WARNING openstack [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.150522] env[65385]: WARNING openstack [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.249204] env[65385]: DEBUG nova.network.neutron [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updated VIF entry in instance network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1107.249520] env[65385]: DEBUG nova.network.neutron [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd731def-a6", "ovs_interfaceid": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1107.356550] env[65385]: DEBUG oslo_concurrency.lockutils [req-577e0049-1655-4da8-8219-58e48d987df0 req-fb29193f-7672-4cf8-b1a2-4c835d832e17 service nova] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1107.369360] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454439, 'name': ReconfigVM_Task, 'duration_secs': 0.299456} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.369360] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 805528b8-96db-4bb8-ba05-2b6332c54624/805528b8-96db-4bb8-ba05-2b6332c54624.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.369714] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance '805528b8-96db-4bb8-ba05-2b6332c54624' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1107.406025] env[65385]: WARNING neutronclient.v2_0.client [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.406598] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.406944] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.486590] env[65385]: INFO nova.compute.manager [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Took 17.46 seconds to build instance. [ 1107.537970] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.538493] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.648697] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.649382] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.660375] env[65385]: WARNING neutronclient.v2_0.client [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.661030] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.661423] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.689204] env[65385]: DEBUG nova.network.neutron [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1107.753222] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.753713] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.760902] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a7d0f96-8d8e-47e9-b673-00b582c52182 req-af0d8a6a-f33b-4634-b8cc-decd69967551 service nova] Releasing lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1107.800114] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updated VIF entry in instance network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1107.800114] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1107.801960] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1107.802601] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43ec5547-77bc-4718-bcf3-fa25318dfd88 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.810924] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1107.810924] env[65385]: value = "task-4454440" [ 1107.810924] env[65385]: _type = "Task" [ 1107.810924] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.826133] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454440, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.859467] env[65385]: WARNING neutronclient.v2_0.client [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.860220] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.860594] env[65385]: WARNING openstack [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.877404] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223d7e8c-5a65-45c8-84f2-60d7a741a32e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.900049] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1639a56f-907f-435e-ba2d-60509f62f3dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.928207] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance '805528b8-96db-4bb8-ba05-2b6332c54624' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1107.978274] env[65385]: DEBUG nova.network.neutron [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updating instance_info_cache with network_info: [{"id": "2341618c-8b69-4d79-8e10-8bec1c24a786", "address": "fa:16:3e:5a:2f:4a", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2341618c-8b", "ovs_interfaceid": "2341618c-8b69-4d79-8e10-8bec1c24a786", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1107.995116] env[65385]: DEBUG oslo_concurrency.lockutils [None req-53e865a7-3b81-409e-b39a-87737e16aad4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.981s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.214660] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53105445-2021-41c1-8c45-0fc5dc9321c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.223717] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b365ca3c-7b67-4210-9732-9e8d9421d26f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.261591] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad1b17d-ac07-43f9-9ca4-d1a65769f7ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.271105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672cd6b3-145b-4994-a976-fc555cb4bf73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.287540] env[65385]: DEBUG nova.compute.provider_tree [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.303873] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.304191] env[65385]: DEBUG nova.compute.manager [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1108.304428] env[65385]: DEBUG nova.compute.manager [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing instance network info cache due to event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1108.304656] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Acquiring lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.304798] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Acquired lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.304957] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1108.321938] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454440, 'name': PowerOffVM_Task, 'duration_secs': 0.301648} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.322261] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1108.323065] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.323443] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2babe60c-26ae-4577-ac33-dfb411ebd730 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.332638] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1108.332638] env[65385]: value = "task-4454441" [ 1108.332638] env[65385]: _type = "Task" [ 1108.332638] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.343667] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1108.344163] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1108.344163] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871157', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'name': 'volume-aee71a50-6144-4280-9b67-ff58bba61633', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b1abf6ab-a3d5-4078-9568-68db0a7f03cc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'serial': 'aee71a50-6144-4280-9b67-ff58bba61633'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1108.345379] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cf56c8-43d7-4c7d-8594-f47c79288824 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.368762] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1aadd7-5cfc-4acb-9d89-53ec192dc791 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.377340] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c7209c-1975-47db-9004-e2f26b74719a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.384594] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "51796586-5084-4656-8a04-e01e7b0847cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.384863] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "51796586-5084-4656-8a04-e01e7b0847cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.402523] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96684157-d0ab-4370-8d31-b556c514338f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.419825] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] The volume has not been displaced from its original location: [datastore1] volume-aee71a50-6144-4280-9b67-ff58bba61633/volume-aee71a50-6144-4280-9b67-ff58bba61633.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1108.424928] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1108.425287] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4644784-63d5-463c-944a-17fcae78eb2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.440358] env[65385]: WARNING neutronclient.v2_0.client [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.440699] env[65385]: WARNING neutronclient.v2_0.client [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.449796] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1108.449796] env[65385]: value = "task-4454442" [ 1108.449796] env[65385]: _type = "Task" [ 1108.449796] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.459277] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454442, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.474797] env[65385]: DEBUG nova.network.neutron [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Port ed7c17c7-54b1-4cee-9cb2-139730bbf715 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1108.481901] env[65385]: INFO nova.compute.manager [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Swapping old allocation on dict_keys(['1af23b69-5ce6-4d6c-8591-1b95ecca8a6b']) held by migration d3c745d8-6882-4477-bdda-3c679cb5ee1b for instance [ 1108.484700] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.485155] env[65385]: DEBUG nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Instance network_info: |[{"id": "2341618c-8b69-4d79-8e10-8bec1c24a786", "address": "fa:16:3e:5a:2f:4a", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2341618c-8b", "ovs_interfaceid": "2341618c-8b69-4d79-8e10-8bec1c24a786", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1108.485723] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:2f:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2341618c-8b69-4d79-8e10-8bec1c24a786', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.494202] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1108.495181] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1108.495424] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3c40604-c003-46df-9e44-8632256aad7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.521900] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1108.521900] env[65385]: value = "task-4454443" [ 1108.521900] env[65385]: _type = "Task" [ 1108.521900] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.527513] env[65385]: DEBUG nova.scheduler.client.report [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Overwriting current allocation {'allocations': {'1af23b69-5ce6-4d6c-8591-1b95ecca8a6b': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 147}}, 'project_id': '1598307c304d46a884a88b3eb5f5d7dc', 'user_id': 'ac74a729b62042eda193030d716719e9', 'consumer_generation': 1} on consumer 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 {{(pid=65385) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1108.536419] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454443, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.594686] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.642936] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.643144] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.643321] env[65385]: DEBUG nova.network.neutron [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1108.790395] env[65385]: DEBUG nova.scheduler.client.report [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1108.808510] env[65385]: WARNING neutronclient.v2_0.client [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.809300] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.809733] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.887694] env[65385]: DEBUG nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1108.928186] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.928767] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.949310] env[65385]: DEBUG nova.compute.manager [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Received event network-vif-plugged-2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1108.949547] env[65385]: DEBUG oslo_concurrency.lockutils [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.949813] env[65385]: DEBUG oslo_concurrency.lockutils [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Lock "b92bc73a-9666-435e-b387-18df7c838f9b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.949903] env[65385]: DEBUG oslo_concurrency.lockutils [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Lock "b92bc73a-9666-435e-b387-18df7c838f9b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.950078] env[65385]: DEBUG nova.compute.manager [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] No waiting events found dispatching network-vif-plugged-2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1108.950278] env[65385]: WARNING nova.compute.manager [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Received unexpected event network-vif-plugged-2341618c-8b69-4d79-8e10-8bec1c24a786 for instance with vm_state building and task_state spawning. [ 1108.950399] env[65385]: DEBUG nova.compute.manager [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Received event network-changed-2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1108.950537] env[65385]: DEBUG nova.compute.manager [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Refreshing instance network info cache due to event network-changed-2341618c-8b69-4d79-8e10-8bec1c24a786. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1108.950708] env[65385]: DEBUG oslo_concurrency.lockutils [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Acquiring lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.950887] env[65385]: DEBUG oslo_concurrency.lockutils [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Acquired lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.950993] env[65385]: DEBUG nova.network.neutron [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Refreshing network info cache for port 2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1108.960864] env[65385]: DEBUG nova.compute.manager [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1108.960864] env[65385]: DEBUG nova.compute.manager [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing instance network info cache due to event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1108.961908] env[65385]: DEBUG oslo_concurrency.lockutils [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Acquiring lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.969284] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454442, 'name': ReconfigVM_Task, 'duration_secs': 0.221156} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.970336] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1108.975336] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc5a31df-6774-4f1e-9bd4-ba245acc73ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.986089] env[65385]: WARNING neutronclient.v2_0.client [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.986713] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.987068] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.006316] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1109.006316] env[65385]: value = "task-4454444" [ 1109.006316] env[65385]: _type = "Task" [ 1109.006316] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.020084] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454444, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.032428] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454443, 'name': CreateVM_Task, 'duration_secs': 0.395255} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.032618] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1109.033137] env[65385]: WARNING neutronclient.v2_0.client [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.033504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.033676] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.033993] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1109.034269] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-336ae99b-8df0-4012-87ae-ec5e7cd9d5eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.042624] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1109.042624] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526af319-daa3-0a64-3564-6d6a9ba0def4" [ 1109.042624] env[65385]: _type = "Task" [ 1109.042624] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.052667] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526af319-daa3-0a64-3564-6d6a9ba0def4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.079404] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updated VIF entry in instance network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1109.080339] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [{"id": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "address": "fa:16:3e:49:27:54", "network": {"id": "380eb239-46a9-4bb2-be3e-66bb113acf76", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1139908934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0a282a49e5f44dc781471d9c2103ca8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38f5da-81", "ovs_interfaceid": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.146525] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.147425] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.147804] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.290470] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.290924] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.298670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.299258] env[65385]: DEBUG nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1109.320665] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.321153] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.321518] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.321828] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.322119] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.334563] env[65385]: INFO nova.compute.manager [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Terminating instance [ 1109.372399] env[65385]: WARNING neutronclient.v2_0.client [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.373550] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.374178] env[65385]: WARNING openstack [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.409090] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.409398] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.410996] env[65385]: INFO nova.compute.claims [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1109.453945] env[65385]: WARNING neutronclient.v2_0.client [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.454678] env[65385]: WARNING openstack [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.455071] env[65385]: WARNING openstack [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.471403] env[65385]: DEBUG nova.network.neutron [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [{"id": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "address": "fa:16:3e:cc:f7:81", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapace973a7-f2", "ovs_interfaceid": "ace973a7-f2da-4905-99cb-e72ce9ef3f18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.519267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.519267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.519267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.526695] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454444, 'name': ReconfigVM_Task, 'duration_secs': 0.190876} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.526993] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871157', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'name': 'volume-aee71a50-6144-4280-9b67-ff58bba61633', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b1abf6ab-a3d5-4078-9568-68db0a7f03cc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aee71a50-6144-4280-9b67-ff58bba61633', 'serial': 'aee71a50-6144-4280-9b67-ff58bba61633'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1109.527296] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.528394] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ed2417-0703-4155-9093-0b6009fcbd80 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.537934] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.543580] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52b940fd-bc67-43bd-9c10-0c3f3a4a8cfa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.557869] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526af319-daa3-0a64-3564-6d6a9ba0def4, 'name': SearchDatastore_Task, 'duration_secs': 0.012522} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.558185] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.558440] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1109.558712] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.558832] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.559037] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1109.559330] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9c6f411-a580-47ee-9151-389613bc10f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.567860] env[65385]: WARNING openstack [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.568305] env[65385]: WARNING openstack [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.576579] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1109.576768] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1109.577576] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c952674-b419-4e2b-8209-f71da036e357 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.583051] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Releasing lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.583359] env[65385]: DEBUG nova.compute.manager [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1109.583529] env[65385]: DEBUG nova.compute.manager [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing instance network info cache due to event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1109.583824] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Acquiring lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.583974] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Acquired lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.584394] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1109.587382] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1109.587382] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522c5e43-d80b-e5e0-1157-e05520fe64e4" [ 1109.587382] env[65385]: _type = "Task" [ 1109.587382] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.598873] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522c5e43-d80b-e5e0-1157-e05520fe64e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.631234] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.631515] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.631747] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Deleting the datastore file [datastore1] b1abf6ab-a3d5-4078-9568-68db0a7f03cc {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.632148] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae9c4c17-36e9-45a8-8ce2-d31abe3d1b18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.640368] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for the task: (returnval){ [ 1109.640368] env[65385]: value = "task-4454446" [ 1109.640368] env[65385]: _type = "Task" [ 1109.640368] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.650315] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.655980] env[65385]: WARNING neutronclient.v2_0.client [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.656681] env[65385]: WARNING openstack [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.657177] env[65385]: WARNING openstack [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.751645] env[65385]: DEBUG nova.network.neutron [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updated VIF entry in instance network info cache for port 2341618c-8b69-4d79-8e10-8bec1c24a786. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1109.752995] env[65385]: DEBUG nova.network.neutron [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updating instance_info_cache with network_info: [{"id": "2341618c-8b69-4d79-8e10-8bec1c24a786", "address": "fa:16:3e:5a:2f:4a", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2341618c-8b", "ovs_interfaceid": "2341618c-8b69-4d79-8e10-8bec1c24a786", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.804957] env[65385]: DEBUG nova.compute.utils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1109.806574] env[65385]: DEBUG nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1109.806810] env[65385]: DEBUG nova.network.neutron [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1109.807152] env[65385]: WARNING neutronclient.v2_0.client [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.807444] env[65385]: WARNING neutronclient.v2_0.client [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.808034] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.808490] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.840368] env[65385]: DEBUG nova.compute.manager [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1109.840651] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.841628] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2722d4a7-3642-4c4c-a58f-c0639eb28149 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.852354] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1109.852354] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-424f54a0-326d-479b-9e5c-3842165af752 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.860474] env[65385]: DEBUG oslo_vmware.api [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1109.860474] env[65385]: value = "task-4454447" [ 1109.860474] env[65385]: _type = "Task" [ 1109.860474] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.868696] env[65385]: DEBUG nova.policy [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b450c500f84f2fa682d70698c398b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d45ff3446a84ef2a1945f7a66b747e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1109.877223] env[65385]: DEBUG oslo_vmware.api [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454447, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.974262] env[65385]: DEBUG oslo_concurrency.lockutils [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-29a449d0-f62d-4430-96ab-9ac43e5a7ad2" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.975342] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10589ed-f174-4ec2-813b-31a00e61faba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.984351] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabc36a3-4452-445e-9783-6e6f0cf36328 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.089940] env[65385]: WARNING neutronclient.v2_0.client [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.090542] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.091374] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.111912] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522c5e43-d80b-e5e0-1157-e05520fe64e4, 'name': SearchDatastore_Task, 'duration_secs': 0.027271} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.112755] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7428fffd-d3af-4979-9b1d-8163e5287283 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.120325] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1110.120325] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ab173d-1578-f4bd-c1e9-b46842fdbefd" [ 1110.120325] env[65385]: _type = "Task" [ 1110.120325] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.131620] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ab173d-1578-f4bd-c1e9-b46842fdbefd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.152904] env[65385]: DEBUG oslo_vmware.api [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Task: {'id': task-4454446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214511} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.153094] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1110.153828] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1110.153828] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1110.210167] env[65385]: DEBUG nova.network.neutron [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Successfully created port: e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1110.238555] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1110.239035] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e5b6b55-e45b-47bd-8ad7-3f49db7bc589 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.249706] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52caee4b-b3ed-4ba2-884e-607eb984135b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.264175] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.264591] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.271648] env[65385]: DEBUG oslo_concurrency.lockutils [req-0aac7c66-9f6a-4dc4-806d-c2ff19879e05 req-808865ea-88d9-4b22-995d-37b1d15a92da service nova] Releasing lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.309998] env[65385]: ERROR nova.compute.manager [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Failed to detach volume aee71a50-6144-4280-9b67-ff58bba61633 from /dev/sda: nova.exception.InstanceNotFound: Instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc could not be found. [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Traceback (most recent call last): [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 4224, in _do_rebuild_instance [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self.driver.rebuild(**kwargs) [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] raise NotImplementedError() [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] NotImplementedError [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] During handling of the above exception, another exception occurred: [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Traceback (most recent call last): [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 3647, in _detach_root_volume [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self.driver.detach_volume(context, old_connection_info, [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] return self._volumeops.detach_volume(connection_info, instance) [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self._detach_volume_vmdk(connection_info, instance) [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] stable_ref.fetch_moref(session) [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] nova.exception.InstanceNotFound: Instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc could not be found. [ 1110.309998] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.318186] env[65385]: DEBUG nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1110.367365] env[65385]: WARNING neutronclient.v2_0.client [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.368024] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.368390] env[65385]: WARNING openstack [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.380998] env[65385]: DEBUG oslo_vmware.api [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454447, 'name': PowerOffVM_Task, 'duration_secs': 0.472673} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.381295] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1110.381589] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1110.382040] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d794deec-a51f-4243-9c58-6c6865fa892b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.479991] env[65385]: DEBUG nova.compute.utils [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Build of instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc aborted: Failed to rebuild volume backed instance. {{(pid=65385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1110.482880] env[65385]: ERROR nova.compute.manager [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc aborted: Failed to rebuild volume backed instance. [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Traceback (most recent call last): [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 4224, in _do_rebuild_instance [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self.driver.rebuild(**kwargs) [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] raise NotImplementedError() [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] NotImplementedError [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] During handling of the above exception, another exception occurred: [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Traceback (most recent call last): [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 3682, in _rebuild_volume_backed_instance [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self._detach_root_volume(context, instance, root_bdm) [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 3661, in _detach_root_volume [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] with excutils.save_and_reraise_exception(): [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self.force_reraise() [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] raise self.value [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 3647, in _detach_root_volume [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self.driver.detach_volume(context, old_connection_info, [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] return self._volumeops.detach_volume(connection_info, instance) [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self._detach_volume_vmdk(connection_info, instance) [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] stable_ref.fetch_moref(session) [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] nova.exception.InstanceNotFound: Instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc could not be found. [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] During handling of the above exception, another exception occurred: [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Traceback (most recent call last): [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 11553, in _error_out_instance_on_exception [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] yield [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 3950, in rebuild_instance [ 1110.482880] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self._do_rebuild_instance_with_claim( [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 4036, in _do_rebuild_instance_with_claim [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self._do_rebuild_instance( [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 4228, in _do_rebuild_instance [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self._rebuild_default_impl(**kwargs) [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 3805, in _rebuild_default_impl [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] self._rebuild_volume_backed_instance( [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] File "/opt/stack/nova/nova/compute/manager.py", line 3697, in _rebuild_volume_backed_instance [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] raise exception.BuildAbortException( [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] nova.exception.BuildAbortException: Build of instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc aborted: Failed to rebuild volume backed instance. [ 1110.483860] env[65385]: ERROR nova.compute.manager [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] [ 1110.488180] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1110.488534] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1110.488870] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Deleting the datastore file [datastore1] f0f981dd-2c6b-4b3a-b62f-4b295fe3457c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1110.492036] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c9fed9a-73a3-46ed-8fac-b35aaa4c92fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.500250] env[65385]: DEBUG oslo_vmware.api [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for the task: (returnval){ [ 1110.500250] env[65385]: value = "task-4454449" [ 1110.500250] env[65385]: _type = "Task" [ 1110.500250] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.504115] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updated VIF entry in instance network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1110.504481] env[65385]: DEBUG nova.network.neutron [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [{"id": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "address": "fa:16:3e:49:27:54", "network": {"id": "380eb239-46a9-4bb2-be3e-66bb113acf76", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1139908934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0a282a49e5f44dc781471d9c2103ca8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38f5da-81", "ovs_interfaceid": "bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1110.521663] env[65385]: DEBUG oslo_vmware.api [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454449, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.529074] env[65385]: WARNING neutronclient.v2_0.client [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.606243] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.606243] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1110.606381] env[65385]: DEBUG nova.network.neutron [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1110.633630] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ab173d-1578-f4bd-c1e9-b46842fdbefd, 'name': SearchDatastore_Task, 'duration_secs': 0.013535} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.634238] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.634504] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b92bc73a-9666-435e-b387-18df7c838f9b/b92bc73a-9666-435e-b387-18df7c838f9b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1110.634793] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8660ca00-8643-4024-bb39-72c3ba0daea8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.648890] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1110.648890] env[65385]: value = "task-4454450" [ 1110.648890] env[65385]: _type = "Task" [ 1110.648890] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.657731] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454450, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.696180] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c99c1a-a7e6-45fb-bb96-a339f6d23544 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.705558] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9ecd3a-3ae3-4902-869e-da17241d7ed8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.750455] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a60d05-e574-4dec-87b0-d1c66ffed9bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.763098] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2df920-d30f-4660-95ac-1a2ea5d8b242 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.785025] env[65385]: DEBUG nova.compute.provider_tree [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.013755] env[65385]: DEBUG oslo_concurrency.lockutils [req-b3a1612c-5508-48af-8107-5e1880031493 req-fcc22660-fa10-40d8-acca-5779eb9b3cf1 service nova] Releasing lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.018204] env[65385]: DEBUG oslo_concurrency.lockutils [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Acquired lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.018496] env[65385]: DEBUG nova.network.neutron [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1111.020183] env[65385]: DEBUG oslo_vmware.api [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Task: {'id': task-4454449, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280396} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.020667] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1111.020851] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1111.021035] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1111.021206] env[65385]: INFO nova.compute.manager [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1111.021569] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1111.021820] env[65385]: DEBUG nova.compute.manager [-] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1111.021950] env[65385]: DEBUG nova.network.neutron [-] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1111.022276] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.023306] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.023492] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.067251] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.095397] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.095731] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57ee40c5-9c0f-4c14-984f-5a24bd3f3cc5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.106161] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1111.106161] env[65385]: value = "task-4454451" [ 1111.106161] env[65385]: _type = "Task" [ 1111.106161] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.111374] env[65385]: WARNING neutronclient.v2_0.client [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.112291] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.113124] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.133896] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.160666] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454450, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.288268] env[65385]: DEBUG nova.scheduler.client.report [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1111.333078] env[65385]: DEBUG nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1111.365955] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1111.366233] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1111.367022] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1111.367415] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1111.367415] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1111.367511] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1111.367700] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1111.367878] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1111.368064] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1111.368229] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1111.368396] env[65385]: DEBUG nova.virt.hardware [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1111.369651] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fe4ce9-7056-42ee-89fd-d22d0131e535 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.380070] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa5a93a-430f-42ad-99a5-bae3a62fad9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.399367] env[65385]: DEBUG nova.compute.manager [req-27435d02-55b4-41c3-a720-32680675e998 req-edbfb076-8709-438e-a597-b66ed19c8165 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received event network-vif-deleted-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1111.399580] env[65385]: INFO nova.compute.manager [req-27435d02-55b4-41c3-a720-32680675e998 req-edbfb076-8709-438e-a597-b66ed19c8165 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Neutron deleted interface bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8; detaching it from the instance and deleting it from the info cache [ 1111.399743] env[65385]: DEBUG nova.network.neutron [req-27435d02-55b4-41c3-a720-32680675e998 req-edbfb076-8709-438e-a597-b66ed19c8165 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1111.504142] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.504555] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.521980] env[65385]: WARNING neutronclient.v2_0.client [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.522739] env[65385]: WARNING openstack [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.523082] env[65385]: WARNING openstack [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.567537] env[65385]: INFO nova.network.neutron [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1111.567799] env[65385]: DEBUG nova.network.neutron [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1111.593648] env[65385]: WARNING neutronclient.v2_0.client [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.594415] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.594823] env[65385]: WARNING openstack [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.616725] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454451, 'name': PowerOffVM_Task, 'duration_secs': 0.262019} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.617182] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.618140] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1111.618411] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1111.618645] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1111.618900] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1111.619112] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1111.619274] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1111.619485] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1111.619637] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1111.619799] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1111.619974] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1111.620217] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1111.625705] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b67f9c1d-7044-4e05-a84e-13d20c57f59d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.643750] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1111.643750] env[65385]: value = "task-4454452" [ 1111.643750] env[65385]: _type = "Task" [ 1111.643750] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.655110] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454452, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.661846] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454450, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.671057} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.662612] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] b92bc73a-9666-435e-b387-18df7c838f9b/b92bc73a-9666-435e-b387-18df7c838f9b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1111.662612] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.662612] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01d88fe5-12e2-4e0c-a31d-f4ce4e8dbbf9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.670576] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1111.670576] env[65385]: value = "task-4454453" [ 1111.670576] env[65385]: _type = "Task" [ 1111.670576] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.687957] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.728459] env[65385]: DEBUG nova.compute.manager [req-6edff1d0-5f44-47b3-b05e-a1625ac0b7a8 req-af79fc67-60ea-4be9-83f2-f240ecf97d7e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Received event network-vif-plugged-e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1111.728607] env[65385]: DEBUG oslo_concurrency.lockutils [req-6edff1d0-5f44-47b3-b05e-a1625ac0b7a8 req-af79fc67-60ea-4be9-83f2-f240ecf97d7e service nova] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.728979] env[65385]: DEBUG oslo_concurrency.lockutils [req-6edff1d0-5f44-47b3-b05e-a1625ac0b7a8 req-af79fc67-60ea-4be9-83f2-f240ecf97d7e service nova] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.728979] env[65385]: DEBUG oslo_concurrency.lockutils [req-6edff1d0-5f44-47b3-b05e-a1625ac0b7a8 req-af79fc67-60ea-4be9-83f2-f240ecf97d7e service nova] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.729194] env[65385]: DEBUG nova.compute.manager [req-6edff1d0-5f44-47b3-b05e-a1625ac0b7a8 req-af79fc67-60ea-4be9-83f2-f240ecf97d7e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] No waiting events found dispatching network-vif-plugged-e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1111.729392] env[65385]: WARNING nova.compute.manager [req-6edff1d0-5f44-47b3-b05e-a1625ac0b7a8 req-af79fc67-60ea-4be9-83f2-f240ecf97d7e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Received unexpected event network-vif-plugged-e5e55898-473b-4cdf-9d34-0b704e923915 for instance with vm_state building and task_state spawning. [ 1111.735135] env[65385]: DEBUG nova.network.neutron [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance_info_cache with network_info: [{"id": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "address": "fa:16:3e:47:9a:0a", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped7c17c7-54", "ovs_interfaceid": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1111.796033] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.797352] env[65385]: DEBUG nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1111.818472] env[65385]: DEBUG nova.network.neutron [-] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1111.822412] env[65385]: DEBUG nova.network.neutron [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Successfully updated port: e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1111.905341] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c94f7ba6-733f-41d5-a1d6-2ffb33468c92 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.915784] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce61469-f1be-4443-8b4e-15bb8bfd3140 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.949075] env[65385]: DEBUG nova.compute.manager [req-27435d02-55b4-41c3-a720-32680675e998 req-edbfb076-8709-438e-a597-b66ed19c8165 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Detach interface failed, port_id=bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8, reason: Instance f0f981dd-2c6b-4b3a-b62f-4b295fe3457c could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1112.070316] env[65385]: DEBUG oslo_concurrency.lockutils [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Releasing lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1112.070316] env[65385]: DEBUG nova.compute.manager [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Received event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1112.070765] env[65385]: DEBUG nova.compute.manager [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing instance network info cache due to event network-changed-bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1112.070765] env[65385]: DEBUG oslo_concurrency.lockutils [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Acquiring lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.070843] env[65385]: DEBUG oslo_concurrency.lockutils [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Acquired lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.071103] env[65385]: DEBUG nova.network.neutron [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Refreshing network info cache for port bc38f5da-81c5-4c40-9f57-e9b5bfbac7c8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1112.154667] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454452, 'name': ReconfigVM_Task, 'duration_secs': 0.189381} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.155604] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3232474a-1e80-4341-9a97-638c8205f8c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.177467] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1112.177742] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1112.177906] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1112.178115] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1112.178268] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1112.178410] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1112.178614] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.178771] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1112.178931] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1112.179101] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1112.179274] env[65385]: DEBUG nova.virt.hardware [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1112.182938] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ead95f19-e556-46a0-b2e6-a4a1dc053467 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.189352] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1112.189352] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dae065-c0ef-d781-d8fe-c277dbd56f2f" [ 1112.189352] env[65385]: _type = "Task" [ 1112.189352] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.192253] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081184} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.195268] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.196010] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46c6d97-34a9-4db9-b73b-24fd0447fc3f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.203818] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dae065-c0ef-d781-d8fe-c277dbd56f2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.221404] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] b92bc73a-9666-435e-b387-18df7c838f9b/b92bc73a-9666-435e-b387-18df7c838f9b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.221748] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-434a8791-62f8-49d8-818e-ea2f05f26746 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.237294] env[65385]: DEBUG oslo_concurrency.lockutils [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1112.247510] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1112.247510] env[65385]: value = "task-4454454" [ 1112.247510] env[65385]: _type = "Task" [ 1112.247510] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.259497] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454454, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.303428] env[65385]: DEBUG nova.compute.utils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1112.304960] env[65385]: DEBUG nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1112.305227] env[65385]: DEBUG nova.network.neutron [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1112.305757] env[65385]: WARNING neutronclient.v2_0.client [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.306032] env[65385]: WARNING neutronclient.v2_0.client [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.306610] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.306991] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.320928] env[65385]: INFO nova.compute.manager [-] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Took 1.30 seconds to deallocate network for instance. [ 1112.327548] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.327734] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.327872] env[65385]: DEBUG nova.network.neutron [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1112.352123] env[65385]: DEBUG nova.policy [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662871f375e441eb98d814cf7d7ac5c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bc8afc699e34f059f18d8244980bc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1112.500358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.500740] env[65385]: DEBUG oslo_concurrency.lockutils [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.575974] env[65385]: WARNING neutronclient.v2_0.client [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.576541] env[65385]: WARNING openstack [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.576925] env[65385]: WARNING openstack [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.621429] env[65385]: DEBUG nova.network.neutron [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1112.675057] env[65385]: DEBUG nova.network.neutron [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Successfully created port: 1388879d-e2b5-421a-bf52-dcbcf2e2de00 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1112.702830] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52dae065-c0ef-d781-d8fe-c277dbd56f2f, 'name': SearchDatastore_Task, 'duration_secs': 0.039734} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.711015] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfiguring VM instance instance-00000050 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1112.711557] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e321cd11-950a-4e99-97f0-da17ea47485f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.733836] env[65385]: DEBUG nova.network.neutron [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1112.736459] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1112.736459] env[65385]: value = "task-4454455" [ 1112.736459] env[65385]: _type = "Task" [ 1112.736459] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.748560] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6eb94e6-1fcc-432e-a21b-87c4ec70c77d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.767204] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454455, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.770559] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be788af-bb49-48d1-a36b-0c3d96b8e505 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.778829] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a87ef14-10e5-44b0-86f1-9ef20dffb646 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.781503] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454454, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.830881] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90df70c2-0b97-480b-b5e9-b9575e68cc5d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.834979] env[65385]: DEBUG nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1112.838486] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.839282] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.839658] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.851097] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421f22d9-03ee-412d-b0f0-438e6cf6c5f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.863283] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance '805528b8-96db-4bb8-ba05-2b6332c54624' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1112.874759] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8873ddb9-ffe9-4ae8-9d99-08e71e50ca7d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.896355] env[65385]: DEBUG nova.compute.provider_tree [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.899986] env[65385]: DEBUG nova.network.neutron [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1112.946194] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.946194] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.062974] env[65385]: WARNING neutronclient.v2_0.client [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.062974] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.063764] env[65385]: WARNING openstack [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.194899] env[65385]: DEBUG nova.network.neutron [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updating instance_info_cache with network_info: [{"id": "e5e55898-473b-4cdf-9d34-0b704e923915", "address": "fa:16:3e:7e:92:5a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5e55898-47", "ovs_interfaceid": "e5e55898-473b-4cdf-9d34-0b704e923915", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1113.237781] env[65385]: DEBUG oslo_concurrency.lockutils [req-40975e07-328d-4f2c-9bdb-43e94101d9f0 req-9260e8f1-6392-4e8d-a61f-e4a93451b354 service nova] Releasing lock "refresh_cache-f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.247937] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454455, 'name': ReconfigVM_Task, 'duration_secs': 0.303687} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.248258] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfigured VM instance instance-00000050 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1113.249116] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51e1d48-e73e-4c0c-80bc-fefa98efc9f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.276485] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.279977] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b23d3c5-e1b8-4050-88f3-965b2008564e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.301862] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454454, 'name': ReconfigVM_Task, 'duration_secs': 0.942562} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.303411] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Reconfigured VM instance instance-0000006c to attach disk [datastore2] b92bc73a-9666-435e-b387-18df7c838f9b/b92bc73a-9666-435e-b387-18df7c838f9b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.304223] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1113.304223] env[65385]: value = "task-4454456" [ 1113.304223] env[65385]: _type = "Task" [ 1113.304223] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.304459] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b24eba2-b333-4138-ac67-9fe73f0e788e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.319969] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454456, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.322223] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1113.322223] env[65385]: value = "task-4454457" [ 1113.322223] env[65385]: _type = "Task" [ 1113.322223] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.332176] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454457, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.375795] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1113.376273] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15aed47d-706b-4db3-a32a-38275c2abc93 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.387206] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1113.387206] env[65385]: value = "task-4454458" [ 1113.387206] env[65385]: _type = "Task" [ 1113.387206] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.400717] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454458, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.404590] env[65385]: DEBUG nova.scheduler.client.report [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1113.698640] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquiring lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.698912] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.699166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquiring lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.699464] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.699557] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.701617] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.701976] env[65385]: DEBUG nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Instance network_info: |[{"id": "e5e55898-473b-4cdf-9d34-0b704e923915", "address": "fa:16:3e:7e:92:5a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5e55898-47", "ovs_interfaceid": "e5e55898-473b-4cdf-9d34-0b704e923915", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1113.702589] env[65385]: INFO nova.compute.manager [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Terminating instance [ 1113.704309] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:92:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5e55898-473b-4cdf-9d34-0b704e923915', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1113.713114] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1113.714907] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1113.715435] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-704d5e3b-0c42-4f53-9e7b-b40dacf7662f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.741963] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1113.741963] env[65385]: value = "task-4454459" [ 1113.741963] env[65385]: _type = "Task" [ 1113.741963] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.754089] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454459, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.776766] env[65385]: DEBUG nova.compute.manager [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Received event network-changed-e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1113.776993] env[65385]: DEBUG nova.compute.manager [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Refreshing instance network info cache due to event network-changed-e5e55898-473b-4cdf-9d34-0b704e923915. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1113.777193] env[65385]: DEBUG oslo_concurrency.lockutils [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Acquiring lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.777330] env[65385]: DEBUG oslo_concurrency.lockutils [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Acquired lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.777571] env[65385]: DEBUG nova.network.neutron [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Refreshing network info cache for port e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1113.817462] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454456, 'name': ReconfigVM_Task, 'duration_secs': 0.429593} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.817727] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2/29a449d0-f62d-4430-96ab-9ac43e5a7ad2.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.818613] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f0e600-8bd2-48d7-8590-c80c3d97b163 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.847304] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03f6e68-8e6a-4f65-9490-2a2a48bbe95f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.853829] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454457, 'name': Rename_Task, 'duration_secs': 0.169567} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.854589] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1113.856247] env[65385]: DEBUG nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1113.858737] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c12ab14-9fbb-4718-9829-e489125f7af5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.881398] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7725afea-81d2-44ce-9a27-2c03f7073e8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.886839] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1113.886839] env[65385]: value = "task-4454460" [ 1113.886839] env[65385]: _type = "Task" [ 1113.886839] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.911113] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048b1028-cbae-414a-944f-658ce65a0f0d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.914448] env[65385]: DEBUG oslo_concurrency.lockutils [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.414s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.914886] env[65385]: INFO nova.compute.manager [None req-283697b5-299b-479d-bffb-d4e479f2575f tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Successfully reverted task state from rebuilding on failure for instance. [ 1113.921044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.083s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.921306] env[65385]: DEBUG nova.objects.instance [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lazy-loading 'resources' on Instance uuid f0f981dd-2c6b-4b3a-b62f-4b295fe3457c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.925864] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1113.926125] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1113.926425] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1113.926530] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1113.926628] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1113.926768] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1113.926980] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.927175] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1113.927345] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1113.927499] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1113.927667] env[65385]: DEBUG nova.virt.hardware [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1113.932530] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69525815-c4db-4274-ac72-324d6193eb72 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.935650] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454460, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.942311] env[65385]: DEBUG oslo_vmware.api [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454458, 'name': PowerOnVM_Task, 'duration_secs': 0.457892} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.942697] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1113.943389] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1113.943556] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c5bf4f-9fc5-4266-ae0e-4a5b0d21f2c8 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance '805528b8-96db-4bb8-ba05-2b6332c54624' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1113.947167] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2e9ad2c-8b83-4134-a864-3655eb9989be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.953423] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53b11c4-a7ce-4432-9ace-7090cd814a48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.958821] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1113.958821] env[65385]: value = "task-4454461" [ 1113.958821] env[65385]: _type = "Task" [ 1113.958821] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.976924] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454461, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.220248] env[65385]: DEBUG nova.compute.manager [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1114.220248] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9e4cf0a-6dfa-4106-8a12-9821ec1c5e81 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.229686] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57d3d27-6751-4a6e-b3bd-8adf85ee3439 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.245041] env[65385]: DEBUG nova.network.neutron [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Successfully updated port: 1388879d-e2b5-421a-bf52-dcbcf2e2de00 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1114.262795] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454459, 'name': CreateVM_Task, 'duration_secs': 0.400831} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.263027] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1114.263499] env[65385]: WARNING neutronclient.v2_0.client [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.263986] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.264188] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.264515] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1114.264940] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0110c63d-a67d-4ea9-a15d-e205f27b545b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.281065] env[65385]: WARNING neutronclient.v2_0.client [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.281779] env[65385]: WARNING openstack [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.282148] env[65385]: WARNING openstack [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.289671] env[65385]: WARNING nova.virt.vmwareapi.driver [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc could not be found. [ 1114.290298] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1114.291532] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a2cc867-30ce-47dc-b091-adb1e25405de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.295269] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1114.295269] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52368858-f495-8f63-e43f-4cce8c5f94dd" [ 1114.295269] env[65385]: _type = "Task" [ 1114.295269] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.304288] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d77ed1-f745-48e4-a15e-6458e4b16bdd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.319501] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52368858-f495-8f63-e43f-4cce8c5f94dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.343789] env[65385]: WARNING nova.virt.vmwareapi.vmops [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1abf6ab-a3d5-4078-9568-68db0a7f03cc could not be found. [ 1114.345037] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1114.345037] env[65385]: INFO nova.compute.manager [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Took 0.13 seconds to destroy the instance on the hypervisor. [ 1114.345037] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1114.345037] env[65385]: DEBUG nova.compute.manager [-] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1114.345037] env[65385]: DEBUG nova.network.neutron [-] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1114.345282] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.345810] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.345997] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.385968] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.395339] env[65385]: WARNING openstack [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.395925] env[65385]: WARNING openstack [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.419691] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454460, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.477947] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454461, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.519475] env[65385]: WARNING neutronclient.v2_0.client [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.520286] env[65385]: WARNING openstack [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.520672] env[65385]: WARNING openstack [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.674460] env[65385]: DEBUG nova.network.neutron [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updated VIF entry in instance network info cache for port e5e55898-473b-4cdf-9d34-0b704e923915. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1114.675077] env[65385]: DEBUG nova.network.neutron [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updating instance_info_cache with network_info: [{"id": "e5e55898-473b-4cdf-9d34-0b704e923915", "address": "fa:16:3e:7e:92:5a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5e55898-47", "ovs_interfaceid": "e5e55898-473b-4cdf-9d34-0b704e923915", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.695184] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636873b5-e3ad-4dc6-b512-c5839d0c9557 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.707673] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22053719-790e-4e5b-9785-427dde0ccb12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.745168] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ddd8ba-43c7-4c64-bc62-63c1eecdf69b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.753480] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "refresh_cache-51796586-5084-4656-8a04-e01e7b0847cf" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.753745] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "refresh_cache-51796586-5084-4656-8a04-e01e7b0847cf" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.753989] env[65385]: DEBUG nova.network.neutron [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1114.756916] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a38fc4-ded9-49c7-9f68-ed5198983633 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.773841] env[65385]: DEBUG nova.compute.provider_tree [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.811855] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52368858-f495-8f63-e43f-4cce8c5f94dd, 'name': SearchDatastore_Task, 'duration_secs': 0.015902} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.812182] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1114.812414] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1114.812639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.812777] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.812950] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.813239] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e60e287b-3086-486b-b67a-11023bc8c9fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.823664] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.823857] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1114.824706] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97b80ac8-b326-444c-8af2-2f1eb94ae4d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.831782] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1114.831782] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e82862-3a64-2a65-980e-e53ad7de48cb" [ 1114.831782] env[65385]: _type = "Task" [ 1114.831782] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.840691] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e82862-3a64-2a65-980e-e53ad7de48cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.844332] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1114.916732] env[65385]: DEBUG oslo_vmware.api [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454460, 'name': PowerOnVM_Task, 'duration_secs': 0.710821} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.917057] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1114.917377] env[65385]: INFO nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Took 8.85 seconds to spawn the instance on the hypervisor. [ 1114.917506] env[65385]: DEBUG nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1114.918410] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98b0240-35c9-466e-91af-b47a9ba207b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.973509] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454461, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.178232] env[65385]: DEBUG oslo_concurrency.lockutils [req-7208a937-dde8-4b0e-8dca-dcdb2c7f3e7d req-ba31956d-dd46-4de2-8728-4b732bdc28c5 service nova] Releasing lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.264035] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.264516] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.277573] env[65385]: DEBUG nova.scheduler.client.report [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1115.342959] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e82862-3a64-2a65-980e-e53ad7de48cb, 'name': SearchDatastore_Task, 'duration_secs': 0.012439} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.343816] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7317df6-1e2d-43f6-916c-96811721ba56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.349998] env[65385]: DEBUG nova.network.neutron [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1115.354931] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1115.354931] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523811e3-ddef-0442-538c-c3e17f87aa52" [ 1115.354931] env[65385]: _type = "Task" [ 1115.354931] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.366660] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]523811e3-ddef-0442-538c-c3e17f87aa52, 'name': SearchDatastore_Task, 'duration_secs': 0.011643} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.367076] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.367407] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1115.367783] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36eb89dc-a73b-447c-b570-d8b006195815 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.378206] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1115.378206] env[65385]: value = "task-4454462" [ 1115.378206] env[65385]: _type = "Task" [ 1115.378206] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.386998] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.393263] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.394085] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.406656] env[65385]: DEBUG nova.network.neutron [-] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.441987] env[65385]: INFO nova.compute.manager [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Took 19.60 seconds to build instance. [ 1115.484859] env[65385]: DEBUG oslo_vmware.api [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454461, 'name': PowerOnVM_Task, 'duration_secs': 1.146785} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.487559] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1115.531269] env[65385]: WARNING neutronclient.v2_0.client [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1115.532040] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.532461] env[65385]: WARNING openstack [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.659090] env[65385]: DEBUG nova.network.neutron [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Updating instance_info_cache with network_info: [{"id": "1388879d-e2b5-421a-bf52-dcbcf2e2de00", "address": "fa:16:3e:54:c0:ca", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1388879d-e2", "ovs_interfaceid": "1388879d-e2b5-421a-bf52-dcbcf2e2de00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.716247] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "805528b8-96db-4bb8-ba05-2b6332c54624" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.716598] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.716895] env[65385]: DEBUG nova.compute.manager [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Going to confirm migration 7 {{(pid=65385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1115.782504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1115.818185] env[65385]: INFO nova.scheduler.client.report [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Deleted allocations for instance f0f981dd-2c6b-4b3a-b62f-4b295fe3457c [ 1115.823384] env[65385]: DEBUG nova.compute.manager [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Received event network-vif-plugged-1388879d-e2b5-421a-bf52-dcbcf2e2de00 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1115.823622] env[65385]: DEBUG oslo_concurrency.lockutils [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Acquiring lock "51796586-5084-4656-8a04-e01e7b0847cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.823846] env[65385]: DEBUG oslo_concurrency.lockutils [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Lock "51796586-5084-4656-8a04-e01e7b0847cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.824030] env[65385]: DEBUG oslo_concurrency.lockutils [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Lock "51796586-5084-4656-8a04-e01e7b0847cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1115.824230] env[65385]: DEBUG nova.compute.manager [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] No waiting events found dispatching network-vif-plugged-1388879d-e2b5-421a-bf52-dcbcf2e2de00 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1115.824456] env[65385]: WARNING nova.compute.manager [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Received unexpected event network-vif-plugged-1388879d-e2b5-421a-bf52-dcbcf2e2de00 for instance with vm_state building and task_state spawning. [ 1115.824613] env[65385]: DEBUG nova.compute.manager [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Received event network-changed-1388879d-e2b5-421a-bf52-dcbcf2e2de00 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1115.824764] env[65385]: DEBUG nova.compute.manager [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Refreshing instance network info cache due to event network-changed-1388879d-e2b5-421a-bf52-dcbcf2e2de00. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1115.824932] env[65385]: DEBUG oslo_concurrency.lockutils [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Acquiring lock "refresh_cache-51796586-5084-4656-8a04-e01e7b0847cf" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.894325] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454462, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.909632] env[65385]: INFO nova.compute.manager [-] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Took 1.56 seconds to deallocate network for instance. [ 1115.946808] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8c5273e9-0112-4629-aaf3-0bbe89d46208 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.118s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.162532] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "refresh_cache-51796586-5084-4656-8a04-e01e7b0847cf" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.162532] env[65385]: DEBUG nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Instance network_info: |[{"id": "1388879d-e2b5-421a-bf52-dcbcf2e2de00", "address": "fa:16:3e:54:c0:ca", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1388879d-e2", "ovs_interfaceid": "1388879d-e2b5-421a-bf52-dcbcf2e2de00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1116.162851] env[65385]: DEBUG oslo_concurrency.lockutils [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Acquired lock "refresh_cache-51796586-5084-4656-8a04-e01e7b0847cf" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.162851] env[65385]: DEBUG nova.network.neutron [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Refreshing network info cache for port 1388879d-e2b5-421a-bf52-dcbcf2e2de00 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1116.164204] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:c0:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1388879d-e2b5-421a-bf52-dcbcf2e2de00', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1116.175876] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1116.178200] env[65385]: WARNING neutronclient.v2_0.client [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.179293] env[65385]: WARNING openstack [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.180843] env[65385]: WARNING openstack [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.195982] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1116.197146] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a03baa3a-b366-4dbd-99d2-f8659915ec87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.238373] env[65385]: WARNING neutronclient.v2_0.client [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.244864] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1116.244864] env[65385]: value = "task-4454463" [ 1116.244864] env[65385]: _type = "Task" [ 1116.244864] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.255736] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454463, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.333022] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2167d554-6d63-4401-b160-e8fec8cf42f2 tempest-ServerRescueTestJSONUnderV235-1121949326 tempest-ServerRescueTestJSONUnderV235-1121949326-project-member] Lock "f0f981dd-2c6b-4b3a-b62f-4b295fe3457c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.011s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.350006] env[65385]: WARNING neutronclient.v2_0.client [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.350403] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.350560] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.350740] env[65385]: DEBUG nova.network.neutron [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1116.350931] env[65385]: DEBUG nova.objects.instance [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lazy-loading 'info_cache' on Instance uuid 805528b8-96db-4bb8-ba05-2b6332c54624 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.367620] env[65385]: WARNING openstack [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.368022] env[65385]: WARNING openstack [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.389330] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571214} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.389597] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1116.389845] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1116.390076] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fe5a946-2cfb-4ca7-856f-322133ede05e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.399222] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1116.399222] env[65385]: value = "task-4454464" [ 1116.399222] env[65385]: _type = "Task" [ 1116.399222] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.410055] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.467788] env[65385]: INFO nova.compute.manager [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Took 0.56 seconds to detach 1 volumes for instance. [ 1116.469898] env[65385]: DEBUG nova.compute.manager [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Deleting volume: aee71a50-6144-4280-9b67-ff58bba61633 {{(pid=65385) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3320}} [ 1116.555183] env[65385]: INFO nova.compute.manager [None req-63c09d78-8c28-4b03-9fdf-b35aa2fa0818 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance to original state: 'active' [ 1116.756604] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454463, 'name': CreateVM_Task, 'duration_secs': 0.405781} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.756797] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1116.757349] env[65385]: WARNING neutronclient.v2_0.client [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.757714] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.757861] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.758203] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1116.758470] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cf51f83-ec77-472a-a9bb-6b3332e41134 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.764386] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1116.764386] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5258b6b2-10a8-c172-cc0e-dfea99b072f0" [ 1116.764386] env[65385]: _type = "Task" [ 1116.764386] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.774999] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5258b6b2-10a8-c172-cc0e-dfea99b072f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.918754] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073623} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.918754] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1116.918754] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1f4928-847a-4dfb-8617-d8add90dd265 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.944900] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1116.950650] env[65385]: WARNING neutronclient.v2_0.client [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.950650] env[65385]: WARNING openstack [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.951015] env[65385]: WARNING openstack [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.962015] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef20aa14-e1cc-490a-90c6-8daa53f7c435 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.983619] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1116.983619] env[65385]: value = "task-4454466" [ 1116.983619] env[65385]: _type = "Task" [ 1116.983619] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.993634] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454466, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.036178] env[65385]: DEBUG nova.compute.manager [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Received event network-changed-2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1117.036178] env[65385]: DEBUG nova.compute.manager [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Refreshing instance network info cache due to event network-changed-2341618c-8b69-4d79-8e10-8bec1c24a786. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1117.036178] env[65385]: DEBUG oslo_concurrency.lockutils [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Acquiring lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.036178] env[65385]: DEBUG oslo_concurrency.lockutils [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Acquired lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.036178] env[65385]: DEBUG nova.network.neutron [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Refreshing network info cache for port 2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1117.045199] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.045623] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.046016] env[65385]: DEBUG nova.objects.instance [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lazy-loading 'resources' on Instance uuid b1abf6ab-a3d5-4078-9568-68db0a7f03cc {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.132638] env[65385]: DEBUG nova.network.neutron [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Updated VIF entry in instance network info cache for port 1388879d-e2b5-421a-bf52-dcbcf2e2de00. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1117.132638] env[65385]: DEBUG nova.network.neutron [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Updating instance_info_cache with network_info: [{"id": "1388879d-e2b5-421a-bf52-dcbcf2e2de00", "address": "fa:16:3e:54:c0:ca", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1388879d-e2", "ovs_interfaceid": "1388879d-e2b5-421a-bf52-dcbcf2e2de00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1117.277443] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5258b6b2-10a8-c172-cc0e-dfea99b072f0, 'name': SearchDatastore_Task, 'duration_secs': 0.028592} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.277703] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.277965] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1117.278891] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.278891] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.278891] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.278891] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6287a662-a913-40ed-9abc-f81d264b3db4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.289412] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.289615] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1117.290418] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20326779-df90-4245-8970-9cc59a30367c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.298042] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1117.298042] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527c5ed9-4f6e-0856-c8e4-df27d6fc1574" [ 1117.298042] env[65385]: _type = "Task" [ 1117.298042] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.307572] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527c5ed9-4f6e-0856-c8e4-df27d6fc1574, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.346579] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.347257] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.347257] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Cleaning up deleted instances {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11916}} [ 1117.360248] env[65385]: WARNING neutronclient.v2_0.client [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.361132] env[65385]: WARNING openstack [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.363016] env[65385]: WARNING openstack [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.498743] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.539568] env[65385]: WARNING neutronclient.v2_0.client [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.540473] env[65385]: WARNING openstack [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.540843] env[65385]: WARNING openstack [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.560741] env[65385]: WARNING openstack [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.561405] env[65385]: WARNING openstack [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.635604] env[65385]: DEBUG oslo_concurrency.lockutils [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] Releasing lock "refresh_cache-51796586-5084-4656-8a04-e01e7b0847cf" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.636078] env[65385]: DEBUG nova.compute.manager [req-460cf88d-96bc-48fe-a217-e19dba366bdb req-42ee7e5a-09e5-40c4-a046-9ce9f9997847 service nova] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Received event network-vif-deleted-b9948bb7-a208-446d-a45e-2a74affe299b {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1117.714869] env[65385]: WARNING neutronclient.v2_0.client [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.715985] env[65385]: WARNING openstack [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.716563] env[65385]: WARNING openstack [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.748522] env[65385]: WARNING openstack [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.749033] env[65385]: WARNING openstack [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.817375] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527c5ed9-4f6e-0856-c8e4-df27d6fc1574, 'name': SearchDatastore_Task, 'duration_secs': 0.011027} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.818545] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01e6cd24-3d48-488d-8323-c64d2604a446 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.822901] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f6dc6b-bf4a-49fa-ace4-e080bc279f35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.827348] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1117.827348] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528bfbb0-0711-a38a-1fe6-a8f9fa7bf8de" [ 1117.827348] env[65385]: _type = "Task" [ 1117.827348] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.834354] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67697327-e85b-4f6f-9fc8-f76f474f5eea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.842765] env[65385]: DEBUG nova.network.neutron [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance_info_cache with network_info: [{"id": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "address": "fa:16:3e:47:9a:0a", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped7c17c7-54", "ovs_interfaceid": "ed7c17c7-54b1-4cee-9cb2-139730bbf715", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1117.844544] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528bfbb0-0711-a38a-1fe6-a8f9fa7bf8de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.896348] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] There are 63 instances to clean {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11925}} [ 1117.896598] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: f0f981dd-2c6b-4b3a-b62f-4b295fe3457c] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1117.903234] env[65385]: WARNING neutronclient.v2_0.client [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.903234] env[65385]: WARNING openstack [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.903234] env[65385]: WARNING openstack [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.911676] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951c4d37-cc42-4130-a40b-5ec5b22634bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.922742] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d559437-37b7-49c9-8256-752722a2f72b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.941968] env[65385]: DEBUG nova.compute.provider_tree [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.000455] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454466, 'name': ReconfigVM_Task, 'duration_secs': 0.716687} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.000455] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.001078] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1d2bfc0-eb6b-48d4-8597-c27d7ea30dd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.008641] env[65385]: DEBUG nova.network.neutron [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updated VIF entry in instance network info cache for port 2341618c-8b69-4d79-8e10-8bec1c24a786. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1118.009016] env[65385]: DEBUG nova.network.neutron [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updating instance_info_cache with network_info: [{"id": "2341618c-8b69-4d79-8e10-8bec1c24a786", "address": "fa:16:3e:5a:2f:4a", "network": {"id": "794f021d-54b6-45d6-8663-0e464c717fc0", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1002016694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ba61d77cc2d4fe1ba3c03466d5985d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2341618c-8b", "ovs_interfaceid": "2341618c-8b69-4d79-8e10-8bec1c24a786", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1118.011674] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1118.011674] env[65385]: value = "task-4454467" [ 1118.011674] env[65385]: _type = "Task" [ 1118.011674] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.022621] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454467, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.080689] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-900639b6-9b98-436f-aaad-ea8391f67393-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.080980] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-900639b6-9b98-436f-aaad-ea8391f67393-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.081371] env[65385]: DEBUG nova.objects.instance [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'flavor' on Instance uuid 900639b6-9b98-436f-aaad-ea8391f67393 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.340109] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]528bfbb0-0711-a38a-1fe6-a8f9fa7bf8de, 'name': SearchDatastore_Task, 'duration_secs': 0.028223} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.340975] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.340975] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 51796586-5084-4656-8a04-e01e7b0847cf/51796586-5084-4656-8a04-e01e7b0847cf.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1118.340975] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df15bd6a-6425-46e7-9b19-2b45b24b4f43 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.345956] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-805528b8-96db-4bb8-ba05-2b6332c54624" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.346225] env[65385]: DEBUG nova.objects.instance [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lazy-loading 'migration_context' on Instance uuid 805528b8-96db-4bb8-ba05-2b6332c54624 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.350038] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1118.350038] env[65385]: value = "task-4454468" [ 1118.350038] env[65385]: _type = "Task" [ 1118.350038] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.360655] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454468, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.402189] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 04ebe83c-cab5-45e1-9f5f-018481f63912] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1118.444815] env[65385]: DEBUG nova.scheduler.client.report [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1118.513011] env[65385]: DEBUG oslo_concurrency.lockutils [req-47a6601e-fe0b-4c55-930f-b7b0e66dcee4 req-5a8388e2-9e2d-4636-a299-a334adfb6079 service nova] Releasing lock "refresh_cache-b92bc73a-9666-435e-b387-18df7c838f9b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.523805] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454467, 'name': Rename_Task, 'duration_secs': 0.397493} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.524117] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1118.524398] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25b1988d-ecfe-4642-b4cb-04c5666fafa3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.533673] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1118.533673] env[65385]: value = "task-4454469" [ 1118.533673] env[65385]: _type = "Task" [ 1118.533673] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.549872] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.585095] env[65385]: WARNING neutronclient.v2_0.client [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1118.586045] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.586461] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.791785] env[65385]: DEBUG nova.objects.instance [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'pci_requests' on Instance uuid 900639b6-9b98-436f-aaad-ea8391f67393 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.849747] env[65385]: DEBUG nova.objects.base [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Object Instance<805528b8-96db-4bb8-ba05-2b6332c54624> lazy-loaded attributes: info_cache,migration_context {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1118.851110] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7406e8-2acb-411c-9ffa-ad3cf8d66232 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.886396] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49c12633-9c9e-4bbb-af64-d454f8127a5b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.892391] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454468, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.900312] env[65385]: DEBUG oslo_vmware.api [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1118.900312] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a7fc6e-0e86-1ba8-152a-cc2ea919af5b" [ 1118.900312] env[65385]: _type = "Task" [ 1118.900312] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.906834] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5ae074b9-4e82-4345-9bc1-ab8d7c930f1c] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1118.913758] env[65385]: DEBUG oslo_vmware.api [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a7fc6e-0e86-1ba8-152a-cc2ea919af5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.951138] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.044149] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454469, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.294682] env[65385]: DEBUG nova.objects.base [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Object Instance<900639b6-9b98-436f-aaad-ea8391f67393> lazy-loaded attributes: flavor,pci_requests {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1119.294895] env[65385]: DEBUG nova.network.neutron [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1119.295277] env[65385]: WARNING neutronclient.v2_0.client [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.295567] env[65385]: WARNING neutronclient.v2_0.client [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.296213] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.296549] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.367392] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454468, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620039} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.367859] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 51796586-5084-4656-8a04-e01e7b0847cf/51796586-5084-4656-8a04-e01e7b0847cf.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1119.367859] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1119.368095] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00489994-cf7d-4fc4-b551-6e08e28ada5b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.376265] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1119.376265] env[65385]: value = "task-4454470" [ 1119.376265] env[65385]: _type = "Task" [ 1119.376265] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.387864] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.391820] env[65385]: DEBUG nova.policy [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1119.394665] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.394933] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.395173] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.395361] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.395523] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.397934] env[65385]: INFO nova.compute.manager [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Terminating instance [ 1119.416912] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 77d718c8-4959-483a-8717-c00f2be1d0bb] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1119.419844] env[65385]: DEBUG oslo_vmware.api [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a7fc6e-0e86-1ba8-152a-cc2ea919af5b, 'name': SearchDatastore_Task, 'duration_secs': 0.019288} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.421187] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.421753] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.472086] env[65385]: DEBUG oslo_concurrency.lockutils [None req-842ccbcf-86f9-443a-bb3a-6bf5b1c95aa5 tempest-ServerActionsV293TestJSON-2034163916 tempest-ServerActionsV293TestJSON-2034163916-project-member] Lock "b1abf6ab-a3d5-4078-9568-68db0a7f03cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.773s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.546876] env[65385]: DEBUG oslo_vmware.api [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454469, 'name': PowerOnVM_Task, 'duration_secs': 0.751753} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.547174] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1119.547380] env[65385]: INFO nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Took 8.21 seconds to spawn the instance on the hypervisor. [ 1119.547613] env[65385]: DEBUG nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1119.548418] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f2af77-50b1-4b6d-80b4-80e6c4c669c5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.886685] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076726} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.886950] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1119.887738] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efd97cf-958e-47d7-b309-fee94e1f13fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.910191] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 51796586-5084-4656-8a04-e01e7b0847cf/51796586-5084-4656-8a04-e01e7b0847cf.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.911072] env[65385]: DEBUG nova.compute.manager [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1119.911310] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1119.911538] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f2ee841-ce67-4473-a286-8340c513354b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.927243] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d0efb24-4d53-4006-aa08-7747d2cf381d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.928943] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: e13d7fec-250a-412e-b952-f189214c8c1e] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1119.940626] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1119.940626] env[65385]: value = "task-4454472" [ 1119.940626] env[65385]: _type = "Task" [ 1119.940626] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.942185] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1119.942185] env[65385]: value = "task-4454471" [ 1119.942185] env[65385]: _type = "Task" [ 1119.942185] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.957878] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.072397] env[65385]: INFO nova.compute.manager [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Took 13.89 seconds to build instance. [ 1120.161174] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a1fc89-725f-4522-b7aa-3cfb7cf75bd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.170121] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "54faee8c-6b2d-4864-bf96-26b9dfba583b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.170637] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "54faee8c-6b2d-4864-bf96-26b9dfba583b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.175507] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f5dd6c-c152-4a2b-97dc-d0b8b8d1e421 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.213390] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673666c3-c3f8-4fb0-bc8e-abafbb1a64eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.222817] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5db517b-7250-4c65-836a-f89ab0a7a871 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.241292] env[65385]: DEBUG nova.compute.provider_tree [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.433942] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5470afe7-6dd0-4778-9b83-2956ede04b7f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1120.455429] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454472, 'name': ReconfigVM_Task, 'duration_secs': 0.471743} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.458745] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 51796586-5084-4656-8a04-e01e7b0847cf/51796586-5084-4656-8a04-e01e7b0847cf.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.459619] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454471, 'name': PowerOffVM_Task, 'duration_secs': 0.38072} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.459815] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4bc0589d-fb94-4135-bdc6-9f3e23baa00b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.461723] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1120.461941] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1120.462144] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871164', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'name': 'volume-9ecf961e-bf7d-4631-ad04-666b1111075d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '29a449d0-f62d-4430-96ab-9ac43e5a7ad2', 'attached_at': '2025-11-14T16:55:49.000000', 'detached_at': '', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'serial': '9ecf961e-bf7d-4631-ad04-666b1111075d'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1120.462977] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d81b190-5441-4658-a430-d84914d3a024 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.488356] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13f612d-b587-4d24-b3ca-86286666df18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.492086] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1120.492086] env[65385]: value = "task-4454473" [ 1120.492086] env[65385]: _type = "Task" [ 1120.492086] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.503662] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ddb0ed-e960-431d-b5d7-4ce7fc12d637 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.510121] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454473, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.532325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f7f776-9ecc-4b61-b4a6-85fb04017e58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.549641] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The volume has not been displaced from its original location: [datastore1] volume-9ecf961e-bf7d-4631-ad04-666b1111075d/volume-9ecf961e-bf7d-4631-ad04-666b1111075d.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1120.555512] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1120.555925] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a96408d3-4768-4112-bf33-68d6763291a7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.574902] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cfcb5d25-7da1-4de0-8a1f-267a53f60abd tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.406s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.577020] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1120.577020] env[65385]: value = "task-4454474" [ 1120.577020] env[65385]: _type = "Task" [ 1120.577020] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.588041] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454474, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.676276] env[65385]: DEBUG nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1120.744449] env[65385]: DEBUG nova.scheduler.client.report [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1120.801168] env[65385]: DEBUG nova.compute.manager [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Received event network-changed-e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1120.801168] env[65385]: DEBUG nova.compute.manager [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Refreshing instance network info cache due to event network-changed-e5e55898-473b-4cdf-9d34-0b704e923915. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1120.801294] env[65385]: DEBUG oslo_concurrency.lockutils [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Acquiring lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.801330] env[65385]: DEBUG oslo_concurrency.lockutils [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Acquired lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1120.803099] env[65385]: DEBUG nova.network.neutron [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Refreshing network info cache for port e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1120.938296] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 537393de-ea48-4d34-8e04-95331bbc099d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1120.990156] env[65385]: DEBUG nova.compute.manager [req-174384a1-366e-41e7-ae46-cbb292f7e693 req-a2a323d6-3891-4a54-be0c-502ace2c79b1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-vif-plugged-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1120.990376] env[65385]: DEBUG oslo_concurrency.lockutils [req-174384a1-366e-41e7-ae46-cbb292f7e693 req-a2a323d6-3891-4a54-be0c-502ace2c79b1 service nova] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.990660] env[65385]: DEBUG oslo_concurrency.lockutils [req-174384a1-366e-41e7-ae46-cbb292f7e693 req-a2a323d6-3891-4a54-be0c-502ace2c79b1 service nova] Lock "900639b6-9b98-436f-aaad-ea8391f67393-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.990729] env[65385]: DEBUG oslo_concurrency.lockutils [req-174384a1-366e-41e7-ae46-cbb292f7e693 req-a2a323d6-3891-4a54-be0c-502ace2c79b1 service nova] Lock "900639b6-9b98-436f-aaad-ea8391f67393-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.990891] env[65385]: DEBUG nova.compute.manager [req-174384a1-366e-41e7-ae46-cbb292f7e693 req-a2a323d6-3891-4a54-be0c-502ace2c79b1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] No waiting events found dispatching network-vif-plugged-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1120.991071] env[65385]: WARNING nova.compute.manager [req-174384a1-366e-41e7-ae46-cbb292f7e693 req-a2a323d6-3891-4a54-be0c-502ace2c79b1 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received unexpected event network-vif-plugged-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 for instance with vm_state active and task_state None. [ 1121.003731] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454473, 'name': Rename_Task, 'duration_secs': 0.21674} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.005451] env[65385]: DEBUG nova.network.neutron [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Successfully updated port: 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1121.007018] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1121.007193] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b7f80a6-1dc7-48c3-8f0b-0c470242c39b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.016032] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1121.016032] env[65385]: value = "task-4454475" [ 1121.016032] env[65385]: _type = "Task" [ 1121.016032] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.026481] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.087186] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454474, 'name': ReconfigVM_Task, 'duration_secs': 0.449594} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.087850] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1121.094033] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-474e95ce-d795-4707-8d9c-aa024f57c879 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.110655] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1121.110655] env[65385]: value = "task-4454476" [ 1121.110655] env[65385]: _type = "Task" [ 1121.110655] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.123217] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454476, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.203323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.304727] env[65385]: WARNING neutronclient.v2_0.client [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.305521] env[65385]: WARNING openstack [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.306067] env[65385]: WARNING openstack [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.427815] env[65385]: WARNING openstack [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.428360] env[65385]: WARNING openstack [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.442149] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 8a974e99-a7b1-438e-aace-9ae82352495c] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1121.497045] env[65385]: WARNING neutronclient.v2_0.client [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.497764] env[65385]: WARNING openstack [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.498128] env[65385]: WARNING openstack [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.509973] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.509973] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.509973] env[65385]: DEBUG nova.network.neutron [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1121.527275] env[65385]: DEBUG oslo_vmware.api [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454475, 'name': PowerOnVM_Task, 'duration_secs': 0.501285} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.527538] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.527732] env[65385]: INFO nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Took 7.67 seconds to spawn the instance on the hypervisor. [ 1121.527905] env[65385]: DEBUG nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1121.528718] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44dfe563-6883-4dd3-ad20-957d993a7146 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.596824] env[65385]: DEBUG nova.network.neutron [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updated VIF entry in instance network info cache for port e5e55898-473b-4cdf-9d34-0b704e923915. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1121.597321] env[65385]: DEBUG nova.network.neutron [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updating instance_info_cache with network_info: [{"id": "e5e55898-473b-4cdf-9d34-0b704e923915", "address": "fa:16:3e:7e:92:5a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5e55898-47", "ovs_interfaceid": "e5e55898-473b-4cdf-9d34-0b704e923915", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1121.622365] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454476, 'name': ReconfigVM_Task, 'duration_secs': 0.215513} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.622604] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871164', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'name': 'volume-9ecf961e-bf7d-4631-ad04-666b1111075d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '29a449d0-f62d-4430-96ab-9ac43e5a7ad2', 'attached_at': '2025-11-14T16:55:49.000000', 'detached_at': '', 'volume_id': '9ecf961e-bf7d-4631-ad04-666b1111075d', 'serial': '9ecf961e-bf7d-4631-ad04-666b1111075d'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1121.622882] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1121.623712] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b154efe5-be54-4490-83e3-b5eaaa45b29a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.631632] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1121.631893] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a796d004-a7d5-44df-9e0b-2192bae5a5d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.720819] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1121.720819] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1121.721341] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleting the datastore file [datastore2] 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1121.721341] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdef2dae-a78b-44a3-bdf1-d18af3f3f5ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.729902] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1121.729902] env[65385]: value = "task-4454478" [ 1121.729902] env[65385]: _type = "Task" [ 1121.729902] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.741889] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.757021] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.335s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.759996] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.557s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1121.761688] env[65385]: INFO nova.compute.claims [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1121.945493] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 6c0f016c-9fdd-4cee-a2bc-749b7bb0f6d5] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1122.014286] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.014518] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.048822] env[65385]: INFO nova.compute.manager [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Took 12.65 seconds to build instance. [ 1122.075529] env[65385]: WARNING nova.network.neutron [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] 16967e55-c149-4a58-905f-c47b7facac11 already exists in list: networks containing: ['16967e55-c149-4a58-905f-c47b7facac11']. ignoring it [ 1122.100404] env[65385]: DEBUG oslo_concurrency.lockutils [req-72c4bad1-fa1c-4e85-9292-6c4498e5ad6b req-45fcbb38-96b7-41cd-9595-7b1535557d4e service nova] Releasing lock "refresh_cache-62c30ded-c851-477f-b1d4-921268a6ef1c" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1122.113834] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.114263] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.184611] env[65385]: WARNING neutronclient.v2_0.client [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.185411] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.185771] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.244599] env[65385]: DEBUG oslo_vmware.api [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208861} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.244864] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1122.245054] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1122.245235] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1122.245403] env[65385]: INFO nova.compute.manager [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Took 2.33 seconds to destroy the instance on the hypervisor. [ 1122.245641] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1122.245836] env[65385]: DEBUG nova.compute.manager [-] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1122.245963] env[65385]: DEBUG nova.network.neutron [-] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1122.246243] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.246765] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.248036] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.295902] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.296311] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.302983] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.345612] env[65385]: INFO nova.scheduler.client.report [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted allocation for migration 3e37f651-5be2-4236-9a1c-b5bd5eb2e876 [ 1122.449419] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: cb99e6ec-fefc-4f44-86a0-9d886e9c722d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1122.481186] env[65385]: WARNING neutronclient.v2_0.client [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.481925] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.482331] env[65385]: WARNING openstack [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.551326] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a5be460-9558-406f-9199-2c3aca31746e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "51796586-5084-4656-8a04-e01e7b0847cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.166s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.788122] env[65385]: DEBUG nova.network.neutron [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "address": "fa:16:3e:6b:e5:fd", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap330c3a6f-16", "ovs_interfaceid": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1122.854305] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c89fe598-c792-42aa-995a-e3e948de76dd tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.136s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.952764] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: a5fe2ad9-6dfe-485b-bf27-beb218d31467] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1122.990127] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8813f2-3952-4f66-bd4e-5cf22e617049 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.003032] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f62f10-3cb5-4a59-b398-5bb9a0bb3f6f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.040087] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4900937-9c70-4a44-9ea9-2626a432c331 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.044476] env[65385]: DEBUG nova.compute.manager [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-changed-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1123.044748] env[65385]: DEBUG nova.compute.manager [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing instance network info cache due to event network-changed-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1123.044980] env[65385]: DEBUG oslo_concurrency.lockutils [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.052416] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3228afd8-53ea-4cd4-903b-be48de1bafdd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.072460] env[65385]: DEBUG nova.compute.provider_tree [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.204637] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "51796586-5084-4656-8a04-e01e7b0847cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.204934] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "51796586-5084-4656-8a04-e01e7b0847cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.205168] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "51796586-5084-4656-8a04-e01e7b0847cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.205346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "51796586-5084-4656-8a04-e01e7b0847cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.205551] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "51796586-5084-4656-8a04-e01e7b0847cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.208331] env[65385]: INFO nova.compute.manager [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Terminating instance [ 1123.290574] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1123.291321] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.291489] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.293056] env[65385]: DEBUG oslo_concurrency.lockutils [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.293056] env[65385]: DEBUG nova.network.neutron [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing network info cache for port 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1123.293803] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5d865e-46fe-4410-a45f-635b81d4233b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.312300] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1123.312550] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1123.312705] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1123.312940] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1123.313113] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1123.313259] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1123.313463] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.313641] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1123.313814] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1123.313971] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1123.314155] env[65385]: DEBUG nova.virt.hardware [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1123.320763] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Reconfiguring VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1123.322116] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-954737a8-8d3e-4463-8e89-a98e55960260 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.335201] env[65385]: DEBUG nova.network.neutron [-] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1123.344326] env[65385]: DEBUG oslo_vmware.api [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1123.344326] env[65385]: value = "task-4454479" [ 1123.344326] env[65385]: _type = "Task" [ 1123.344326] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.354602] env[65385]: DEBUG oslo_vmware.api [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454479, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.456293] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 7b6bab87-d2a3-4131-896f-3e4fb197ecd6] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1123.575800] env[65385]: DEBUG nova.scheduler.client.report [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1123.712238] env[65385]: DEBUG nova.compute.manager [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1123.712490] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1123.713826] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dfd576-70a1-4140-94be-be3eb1dfea1a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.721884] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1123.722149] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c5f7a92-9e8e-4d00-a1d4-e3dc81b14c3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.729036] env[65385]: DEBUG oslo_vmware.api [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1123.729036] env[65385]: value = "task-4454480" [ 1123.729036] env[65385]: _type = "Task" [ 1123.729036] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.737694] env[65385]: DEBUG oslo_vmware.api [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.798160] env[65385]: WARNING neutronclient.v2_0.client [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1123.798934] env[65385]: WARNING openstack [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.799350] env[65385]: WARNING openstack [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.841021] env[65385]: INFO nova.compute.manager [-] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Took 1.59 seconds to deallocate network for instance. [ 1123.856401] env[65385]: DEBUG oslo_vmware.api [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454479, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.902113] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "805528b8-96db-4bb8-ba05-2b6332c54624" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.903116] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.903200] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.903444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.903822] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.906402] env[65385]: INFO nova.compute.manager [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Terminating instance [ 1123.933763] env[65385]: WARNING openstack [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.935136] env[65385]: WARNING openstack [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.961216] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 167e1b46-af52-4a7a-9964-edabc8155287] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1124.006458] env[65385]: WARNING neutronclient.v2_0.client [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.007188] env[65385]: WARNING openstack [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.007544] env[65385]: WARNING openstack [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.082607] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.083495] env[65385]: DEBUG nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1124.102823] env[65385]: DEBUG nova.network.neutron [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updated VIF entry in instance network info cache for port 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1124.103059] env[65385]: DEBUG nova.network.neutron [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "address": "fa:16:3e:6b:e5:fd", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap330c3a6f-16", "ovs_interfaceid": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.246644] env[65385]: DEBUG oslo_vmware.api [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454480, 'name': PowerOffVM_Task, 'duration_secs': 0.244434} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.247013] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1124.247638] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1124.248721] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f4b0cd9-553d-4672-87dc-8f1796f1494e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.354799] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1124.355061] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1124.355246] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleting the datastore file [datastore1] 51796586-5084-4656-8a04-e01e7b0847cf {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1124.355547] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b470348-8d38-4e23-bce0-aecc0114f528 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.371525] env[65385]: DEBUG oslo_vmware.api [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454479, 'name': ReconfigVM_Task, 'duration_secs': 0.990258} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.372868] env[65385]: WARNING neutronclient.v2_0.client [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.375043] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.375043] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Reconfigured VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1124.380449] env[65385]: DEBUG oslo_vmware.api [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1124.380449] env[65385]: value = "task-4454482" [ 1124.380449] env[65385]: _type = "Task" [ 1124.380449] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.390694] env[65385]: DEBUG oslo_vmware.api [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.404508] env[65385]: INFO nova.compute.manager [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Took 0.56 seconds to detach 1 volumes for instance. [ 1124.416127] env[65385]: DEBUG nova.compute.manager [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1124.416355] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1124.417307] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6740f44-b9ef-4f32-93f2-2a13ef4ea4c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.426707] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1124.428397] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3666d4e0-91f1-42c4-b7a6-5947a03618f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.436866] env[65385]: DEBUG oslo_vmware.api [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1124.436866] env[65385]: value = "task-4454483" [ 1124.436866] env[65385]: _type = "Task" [ 1124.436866] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.447687] env[65385]: DEBUG oslo_vmware.api [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.464159] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: c003dd80-d419-4a5a-8f86-dcac1a1e789c] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1124.560966] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.561278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.589579] env[65385]: DEBUG nova.compute.utils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1124.592088] env[65385]: DEBUG nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 1124.608257] env[65385]: DEBUG oslo_concurrency.lockutils [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.608816] env[65385]: DEBUG nova.compute.manager [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Received event network-vif-deleted-ace973a7-f2da-4905-99cb-e72ce9ef3f18 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1124.608816] env[65385]: INFO nova.compute.manager [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Neutron deleted interface ace973a7-f2da-4905-99cb-e72ce9ef3f18; detaching it from the instance and deleting it from the info cache [ 1124.608946] env[65385]: DEBUG nova.network.neutron [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.882714] env[65385]: DEBUG oslo_concurrency.lockutils [None req-72ed514d-5c92-4485-b975-3e7c38425806 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-900639b6-9b98-436f-aaad-ea8391f67393-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.802s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.902873] env[65385]: DEBUG oslo_vmware.api [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167013} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.904250] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1124.904250] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1124.904250] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1124.904435] env[65385]: INFO nova.compute.manager [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1124.904555] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1124.904808] env[65385]: DEBUG nova.compute.manager [-] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1124.904880] env[65385]: DEBUG nova.network.neutron [-] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1124.905098] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.905633] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.905909] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.914309] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.914620] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.914851] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.947446] env[65385]: DEBUG oslo_vmware.api [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454483, 'name': PowerOffVM_Task, 'duration_secs': 0.319724} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.948551] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.950417] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1124.950597] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1124.951386] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61dca621-f8f5-4785-b4bc-8a99cb3fb1b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.967811] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 55936da6-fcdf-4291-9b3a-23fa8b7a56ef] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1125.021564] env[65385]: INFO nova.scheduler.client.report [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted allocations for instance 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 [ 1125.034943] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1125.035377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1125.035672] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleting the datastore file [datastore2] 805528b8-96db-4bb8-ba05-2b6332c54624 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1125.036106] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f781c998-532b-49af-b529-f4e776b5caf3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.047786] env[65385]: DEBUG oslo_vmware.api [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1125.047786] env[65385]: value = "task-4454485" [ 1125.047786] env[65385]: _type = "Task" [ 1125.047786] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.060786] env[65385]: DEBUG oslo_vmware.api [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454485, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.064579] env[65385]: DEBUG nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1125.094125] env[65385]: DEBUG nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1125.112327] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8c3da2ac-6c16-4f16-a829-a9ea34a438c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.124942] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea65c7ef-f27e-402b-99f5-c7155a030100 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.163771] env[65385]: DEBUG nova.compute.manager [req-5079d52d-f308-4dca-b30d-1d262b7cf3e7 req-497fb9fe-7e2b-4e02-8103-e6a52a41fad3 service nova] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Detach interface failed, port_id=ace973a7-f2da-4905-99cb-e72ce9ef3f18, reason: Instance 29a449d0-f62d-4430-96ab-9ac43e5a7ad2 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1125.316199] env[65385]: DEBUG nova.compute.manager [req-771e987b-d65e-4497-a49e-bf626cdc9664 req-bd66f332-f59c-4712-90db-ef4fed31ecbb service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Received event network-vif-deleted-1388879d-e2b5-421a-bf52-dcbcf2e2de00 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1125.316275] env[65385]: INFO nova.compute.manager [req-771e987b-d65e-4497-a49e-bf626cdc9664 req-bd66f332-f59c-4712-90db-ef4fed31ecbb service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Neutron deleted interface 1388879d-e2b5-421a-bf52-dcbcf2e2de00; detaching it from the instance and deleting it from the info cache [ 1125.316414] env[65385]: DEBUG nova.network.neutron [req-771e987b-d65e-4497-a49e-bf626cdc9664 req-bd66f332-f59c-4712-90db-ef4fed31ecbb service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1125.402665] env[65385]: DEBUG nova.compute.manager [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1125.476023] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: f4af288e-b9c9-4273-8e65-e726ae039d03] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1125.532437] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c8fc1236-ea78-4d63-b420-bfa7d5f51211 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "29a449d0-f62d-4430-96ab-9ac43e5a7ad2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.137s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.562933] env[65385]: DEBUG oslo_vmware.api [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454485, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.436335} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.563266] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1125.563496] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1125.563794] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1125.564025] env[65385]: INFO nova.compute.manager [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1125.564311] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1125.564575] env[65385]: DEBUG nova.compute.manager [-] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1125.564712] env[65385]: DEBUG nova.network.neutron [-] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1125.565013] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.565858] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.566324] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.606693] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.606982] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.608874] env[65385]: INFO nova.compute.claims [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1125.625270] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.778126] env[65385]: DEBUG nova.network.neutron [-] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1125.819765] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e47cd1df-eafc-4893-8145-f08e7a43b021 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.832908] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d49acf-2773-4623-95b1-85563ef06ecc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.870140] env[65385]: DEBUG nova.compute.manager [req-771e987b-d65e-4497-a49e-bf626cdc9664 req-bd66f332-f59c-4712-90db-ef4fed31ecbb service nova] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Detach interface failed, port_id=1388879d-e2b5-421a-bf52-dcbcf2e2de00, reason: Instance 51796586-5084-4656-8a04-e01e7b0847cf could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1125.933432] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.978748] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: f2e5b164-40e5-49f6-b3cc-79fe6e02d0e5] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1126.105892] env[65385]: DEBUG nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1126.140807] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1126.141098] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1126.141257] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1126.141432] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1126.141576] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1126.141716] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1126.141919] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.142128] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1126.142410] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1126.142468] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1126.142595] env[65385]: DEBUG nova.virt.hardware [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1126.144069] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3175d8c9-b7f8-4e05-87ac-1821bc0a87ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.155918] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f35c89a-ebce-450e-82d2-c1f62aa42f0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.174086] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1126.179304] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Creating folder: Project (31eb3435d84e4de5aa47fd2272fa8924). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1126.179697] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5509818e-0123-450b-a6fd-74dc17a7331e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.193163] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Created folder: Project (31eb3435d84e4de5aa47fd2272fa8924) in parent group-v870881. [ 1126.193515] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Creating folder: Instances. Parent ref: group-v871181. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1126.193951] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36cc267f-bc3c-47c5-85f2-070209549bfe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.207914] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Created folder: Instances in parent group-v871181. [ 1126.208197] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1126.208419] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1126.208735] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d3567cb-dde2-4343-80ad-fb5d420c134d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.227984] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1126.227984] env[65385]: value = "task-4454488" [ 1126.227984] env[65385]: _type = "Task" [ 1126.227984] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.237204] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454488, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.280843] env[65385]: INFO nova.compute.manager [-] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Took 1.38 seconds to deallocate network for instance. [ 1126.375655] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-900639b6-9b98-436f-aaad-ea8391f67393-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.376509] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-900639b6-9b98-436f-aaad-ea8391f67393-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.417491] env[65385]: DEBUG nova.network.neutron [-] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1126.484767] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 11decb5b-d07c-4d6a-a4f3-f9e8d1cd2ac3] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1126.740418] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454488, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.781591] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.781988] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.790662] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.848589] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a846c343-6e9e-4ad1-bb09-c709faf21727 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.856736] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79e06cf-4ac0-4dcb-ab13-c910b6850182 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.891924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.892127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1126.893194] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d7c75c-8542-446d-93ae-e0800f82ee9f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.896331] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e4556a-77d2-4eb6-930d-4e3220093606 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.917346] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2319b0-703e-46b6-8eeb-be44f57c46f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.922299] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93522099-bbd7-47c6-84f3-87e1c0fa91c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.925450] env[65385]: INFO nova.compute.manager [-] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Took 1.36 seconds to deallocate network for instance. [ 1126.936415] env[65385]: DEBUG nova.compute.provider_tree [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.957290] env[65385]: WARNING neutronclient.v2_0.client [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.962908] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Reconfiguring VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1126.964219] env[65385]: DEBUG nova.scheduler.client.report [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1126.967838] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09a03c08-fe7b-4c8b-ac92-c95bcfc7432e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.983363] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.376s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.983910] env[65385]: DEBUG nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1126.986608] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.053s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.988981] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 78f9e2e6-e7f9-401c-b5d3-a2e58e8c226b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1126.994387] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1126.994387] env[65385]: value = "task-4454489" [ 1126.994387] env[65385]: _type = "Task" [ 1126.994387] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.004307] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.238672] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454488, 'name': CreateVM_Task, 'duration_secs': 0.57329} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.238877] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1127.239417] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.239590] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1127.239909] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1127.240193] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b04620b0-7141-4ba5-bac3-87f916e85c54 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.246120] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1127.246120] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524695de-6e56-9cde-f3b7-d415359c9db4" [ 1127.246120] env[65385]: _type = "Task" [ 1127.246120] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.254992] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524695de-6e56-9cde-f3b7-d415359c9db4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.285662] env[65385]: DEBUG nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1127.342957] env[65385]: DEBUG nova.compute.manager [req-d9a967b9-d56c-4c0c-b53d-17340eb49190 req-519fb042-2df5-4470-9f27-74cfdfef0ee6 service nova] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Received event network-vif-deleted-ed7c17c7-54b1-4cee-9cb2-139730bbf715 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1127.464680] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.491776] env[65385]: INFO nova.compute.claims [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1127.496550] env[65385]: DEBUG nova.compute.utils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1127.498106] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 354527f0-007d-449f-9e15-48ce1d91876f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1127.499936] env[65385]: DEBUG nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1127.500177] env[65385]: DEBUG nova.network.neutron [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1127.500554] env[65385]: WARNING neutronclient.v2_0.client [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.500891] env[65385]: WARNING neutronclient.v2_0.client [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.501525] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.501970] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.522258] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.570245] env[65385]: DEBUG nova.policy [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c1a9d4194964403a8d11abfad4c65a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94cc5bc221c0455ea760f5022db6bdc4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1127.757400] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524695de-6e56-9cde-f3b7-d415359c9db4, 'name': SearchDatastore_Task, 'duration_secs': 0.011979} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.757772] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.757976] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1127.758261] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.758407] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1127.758588] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1127.758862] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30997424-d658-48af-afb8-e803dee08b64 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.768820] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1127.769013] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1127.769801] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82045d8b-d767-4855-9d57-15d5681b6f51 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.775979] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1127.775979] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524af82b-818c-5aa9-460e-465a36b72e15" [ 1127.775979] env[65385]: _type = "Task" [ 1127.775979] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.785444] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524af82b-818c-5aa9-460e-465a36b72e15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.809302] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.911711] env[65385]: DEBUG nova.network.neutron [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Successfully created port: 7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1128.000110] env[65385]: INFO nova.compute.resource_tracker [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating resource usage from migration 77bd2d4a-6632-4319-9c82-871dbac8ac69 [ 1128.013962] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 3febf6af-699a-4fa1-b079-3790cd1095e5] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1128.015751] env[65385]: DEBUG nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1128.018095] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.202339] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a826f17f-54c8-4074-b28a-80212f31bc1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.210756] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e06050f-8f32-438e-94f8-6abe9b127ff3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.242945] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495e325d-d65e-4ebf-a09b-3ed3a376372a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.251394] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98b6a11-40ce-424c-be11-306ff7c19e35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.265623] env[65385]: DEBUG nova.compute.provider_tree [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.286565] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]524af82b-818c-5aa9-460e-465a36b72e15, 'name': SearchDatastore_Task, 'duration_secs': 0.009482} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.287366] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae8b0eca-ad62-4b2e-8bdc-8d308947abff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.294438] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1128.294438] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521cb061-f659-ec93-cf76-76902a8a1193" [ 1128.294438] env[65385]: _type = "Task" [ 1128.294438] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.303688] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521cb061-f659-ec93-cf76-76902a8a1193, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.513932] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.520044] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: e8f9418e-904c-4e3f-8ea9-91ab0fdd080d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1128.768695] env[65385]: DEBUG nova.scheduler.client.report [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1128.805727] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521cb061-f659-ec93-cf76-76902a8a1193, 'name': SearchDatastore_Task, 'duration_secs': 0.010402} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.805727] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1128.805969] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1128.806126] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c14d274-53aa-4ca8-a90e-e6e815e80933 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.813535] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1128.813535] env[65385]: value = "task-4454490" [ 1128.813535] env[65385]: _type = "Task" [ 1128.813535] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.824270] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.015496] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.026133] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: fbc67b68-64ed-410a-a43a-efa3f55b3031] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1129.030701] env[65385]: DEBUG nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1129.068115] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1129.068347] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1129.068515] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1129.068710] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1129.068855] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1129.068999] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1129.069316] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.069474] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1129.069635] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1129.069792] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1129.069970] env[65385]: DEBUG nova.virt.hardware [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1129.070870] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa5c311-55d4-4c01-b3c5-c16cd7ccbb8b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.082819] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bb0b20-cb86-4498-bebe-fb3183c6f216 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.274068] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.287s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.274068] env[65385]: INFO nova.compute.manager [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Migrating [ 1129.281350] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.491s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.281641] env[65385]: DEBUG nova.objects.instance [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'resources' on Instance uuid 51796586-5084-4656-8a04-e01e7b0847cf {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.329061] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506597} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.329348] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1129.329598] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1129.329879] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-114d475f-e380-4e07-822b-f08f40d3ed9e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.337016] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1129.337016] env[65385]: value = "task-4454491" [ 1129.337016] env[65385]: _type = "Task" [ 1129.337016] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.349681] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.379784] env[65385]: DEBUG nova.compute.manager [req-4432ad6d-5715-40cb-9574-9eb3b7b05d12 req-3513d6fa-fc9a-4943-a121-ff4c2bf37d69 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Received event network-vif-plugged-7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1129.380022] env[65385]: DEBUG oslo_concurrency.lockutils [req-4432ad6d-5715-40cb-9574-9eb3b7b05d12 req-3513d6fa-fc9a-4943-a121-ff4c2bf37d69 service nova] Acquiring lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.380236] env[65385]: DEBUG oslo_concurrency.lockutils [req-4432ad6d-5715-40cb-9574-9eb3b7b05d12 req-3513d6fa-fc9a-4943-a121-ff4c2bf37d69 service nova] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.380400] env[65385]: DEBUG oslo_concurrency.lockutils [req-4432ad6d-5715-40cb-9574-9eb3b7b05d12 req-3513d6fa-fc9a-4943-a121-ff4c2bf37d69 service nova] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.380571] env[65385]: DEBUG nova.compute.manager [req-4432ad6d-5715-40cb-9574-9eb3b7b05d12 req-3513d6fa-fc9a-4943-a121-ff4c2bf37d69 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] No waiting events found dispatching network-vif-plugged-7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1129.380749] env[65385]: WARNING nova.compute.manager [req-4432ad6d-5715-40cb-9574-9eb3b7b05d12 req-3513d6fa-fc9a-4943-a121-ff4c2bf37d69 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Received unexpected event network-vif-plugged-7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 for instance with vm_state building and task_state spawning. [ 1129.478103] env[65385]: DEBUG nova.network.neutron [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Successfully updated port: 7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1129.516697] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.520898] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b64042-6130-40a4-b9af-99e31ffad60b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.528986] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc09d39-036e-4ff7-890e-ce33a1be78ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.532507] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: ac6fa516-d954-466e-b8e0-e12440492049] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1129.566706] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6013fb6-57d3-4907-9e64-3b0da523a108 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.576695] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88189850-b984-4614-a206-97ead45b8f40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.593532] env[65385]: DEBUG nova.compute.provider_tree [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.792084] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.792386] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.792679] env[65385]: DEBUG nova.network.neutron [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1129.850656] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07033} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.850953] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1129.851837] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670ac727-b828-4f91-a4c0-5c25db526eeb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.873070] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1129.873771] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31078fd7-8c19-44db-8836-ce4c15900dcf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.895259] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1129.895259] env[65385]: value = "task-4454492" [ 1129.895259] env[65385]: _type = "Task" [ 1129.895259] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.904928] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454492, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.984396] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.984511] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.984748] env[65385]: DEBUG nova.network.neutron [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1130.015787] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.036019] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 048e7f17-7ac1-4d6d-ad0e-a4dc9fafa8d7] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1130.097933] env[65385]: DEBUG nova.scheduler.client.report [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1130.296519] env[65385]: WARNING neutronclient.v2_0.client [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.297247] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.297563] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.410438] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454492, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.421871] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.422334] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.488430] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.488949] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.497653] env[65385]: WARNING neutronclient.v2_0.client [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.498283] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.498579] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.519433] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.535366] env[65385]: DEBUG nova.network.neutron [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1130.539014] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 4b0b8859-b63e-4740-bf67-5733c75e1c70] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1130.559659] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.560220] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.594631] env[65385]: DEBUG nova.network.neutron [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1130.604804] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.323s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.607446] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.143s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.607723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.609546] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.800s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.611195] env[65385]: INFO nova.compute.claims [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1130.643638] env[65385]: INFO nova.scheduler.client.report [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted allocations for instance 51796586-5084-4656-8a04-e01e7b0847cf [ 1130.653216] env[65385]: INFO nova.scheduler.client.report [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted allocations for instance 805528b8-96db-4bb8-ba05-2b6332c54624 [ 1130.656366] env[65385]: WARNING neutronclient.v2_0.client [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.657224] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.657754] env[65385]: WARNING openstack [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.756532] env[65385]: DEBUG nova.network.neutron [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance_info_cache with network_info: [{"id": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "address": "fa:16:3e:b0:35:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e4dcc22-4c", "ovs_interfaceid": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1130.906435] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454492, 'name': ReconfigVM_Task, 'duration_secs': 0.815652} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.906715] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1130.907429] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b85b10a1-031d-41db-a1c2-acc5b0fb3728 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.914236] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1130.914236] env[65385]: value = "task-4454493" [ 1130.914236] env[65385]: _type = "Task" [ 1130.914236] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.922527] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454493, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.018204] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.042418] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: cf37e10a-14f4-4456-bad0-d7528457658d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1131.098014] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.152549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-5c2beadb-d32b-470d-ae33-e18cd3f816a4 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "51796586-5084-4656-8a04-e01e7b0847cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.947s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.171897] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a0d1628-271e-4a4a-9b10-b31e67c94905 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "805528b8-96db-4bb8-ba05-2b6332c54624" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.269s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.259828] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.260336] env[65385]: DEBUG nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Instance network_info: |[{"id": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "address": "fa:16:3e:b0:35:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e4dcc22-4c", "ovs_interfaceid": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1131.260859] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:35:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e4dcc22-4cf8-46f3-9acd-c5ad6685f818', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.268736] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1131.268906] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1131.269158] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ed63467-18a0-48d8-ad16-5a66550cbf5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.290423] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1131.290423] env[65385]: value = "task-4454494" [ 1131.290423] env[65385]: _type = "Task" [ 1131.290423] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.298937] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454494, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.412681] env[65385]: DEBUG nova.compute.manager [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Received event network-changed-7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1131.412875] env[65385]: DEBUG nova.compute.manager [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Refreshing instance network info cache due to event network-changed-7e4dcc22-4cf8-46f3-9acd-c5ad6685f818. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1131.413104] env[65385]: DEBUG oslo_concurrency.lockutils [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Acquiring lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.413244] env[65385]: DEBUG oslo_concurrency.lockutils [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Acquired lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.413397] env[65385]: DEBUG nova.network.neutron [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Refreshing network info cache for port 7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1131.424973] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454493, 'name': Rename_Task, 'duration_secs': 0.127299} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.425240] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.425492] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44e473dd-59c8-4978-8b65-8ad2dfe729ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.434042] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1131.434042] env[65385]: value = "task-4454495" [ 1131.434042] env[65385]: _type = "Task" [ 1131.434042] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.444505] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454495, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.519628] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.546407] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: fa9a5278-1477-485a-9201-a37187488aef] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1131.808535] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454494, 'name': CreateVM_Task, 'duration_secs': 0.310579} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.808838] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1131.810528] env[65385]: WARNING neutronclient.v2_0.client [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.811043] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.811300] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.811751] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1131.812902] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930fb02b-bf47-4850-93a6-846e9940c9b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.816762] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43b4d1d0-eaa5-45ea-84fc-9115a45d5eb6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.827048] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7fed89-30f6-4962-a9a8-9401fdcde533 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.830465] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1131.830465] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1f3aa-715f-abb7-aa89-0ae0573277ec" [ 1131.830465] env[65385]: _type = "Task" [ 1131.830465] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.861704] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e0b29b-5869-4254-98c9-6557e359e010 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.868873] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b1f3aa-715f-abb7-aa89-0ae0573277ec, 'name': SearchDatastore_Task, 'duration_secs': 0.013394} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.869673] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.869885] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.870134] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.870275] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.870447] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.870723] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2526ee2-9223-4517-b459-30d89eb0ef35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.879320] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688dbc3f-bec9-4726-b5ab-371715123696 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.884944] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.885121] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1131.893291] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc909141-43af-45fd-b732-cbc160ed32cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.895994] env[65385]: DEBUG nova.compute.provider_tree [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.900923] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1131.900923] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5255a7c1-4f4a-ee45-e903-151a569053fa" [ 1131.900923] env[65385]: _type = "Task" [ 1131.900923] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.909931] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5255a7c1-4f4a-ee45-e903-151a569053fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.916814] env[65385]: WARNING neutronclient.v2_0.client [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.916983] env[65385]: WARNING openstack [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.917378] env[65385]: WARNING openstack [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.944506] env[65385]: DEBUG oslo_vmware.api [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454495, 'name': PowerOnVM_Task, 'duration_secs': 0.43939} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.944790] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1131.944989] env[65385]: INFO nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Took 5.84 seconds to spawn the instance on the hypervisor. [ 1131.945178] env[65385]: DEBUG nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1131.945987] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4b6cdd-a0a0-472b-962f-bf50569c1a0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.015859] env[65385]: WARNING openstack [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.016455] env[65385]: WARNING openstack [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.027883] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.053385] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 6dae15ad-a340-47e3-ab6b-95b25043ca27] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1132.086754] env[65385]: WARNING neutronclient.v2_0.client [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.087691] env[65385]: WARNING openstack [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.087806] env[65385]: WARNING openstack [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.179273] env[65385]: DEBUG nova.network.neutron [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updated VIF entry in instance network info cache for port 7e4dcc22-4cf8-46f3-9acd-c5ad6685f818. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1132.179650] env[65385]: DEBUG nova.network.neutron [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance_info_cache with network_info: [{"id": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "address": "fa:16:3e:b0:35:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e4dcc22-4c", "ovs_interfaceid": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1132.277989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.277989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.277989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.277989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.277989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.283474] env[65385]: INFO nova.compute.manager [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Terminating instance [ 1132.400197] env[65385]: DEBUG nova.scheduler.client.report [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1132.413739] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5255a7c1-4f4a-ee45-e903-151a569053fa, 'name': SearchDatastore_Task, 'duration_secs': 0.011677} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.415102] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f8d58e-839f-4da9-aa30-6d4583031a99 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.421317] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1132.421317] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52777dcd-0196-c651-ab51-a64cd5dbff88" [ 1132.421317] env[65385]: _type = "Task" [ 1132.421317] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.431689] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52777dcd-0196-c651-ab51-a64cd5dbff88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.463357] env[65385]: INFO nova.compute.manager [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Took 11.28 seconds to build instance. [ 1132.520490] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.557531] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: e02e3848-1b1e-426b-bc0f-24c8e232a9db] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1132.617052] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26ae552-17c5-49ec-9967-ac1e0ab933e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.637603] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance '6c826914-4c6b-456b-a403-2373d30c803f' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.682658] env[65385]: DEBUG oslo_concurrency.lockutils [req-8488a79b-f187-4ba3-b38c-aad90f68b87d req-81b849c4-e5f0-4f3e-bb41-2604b58cdb61 service nova] Releasing lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.786927] env[65385]: INFO nova.compute.manager [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Rebuilding instance [ 1132.789797] env[65385]: DEBUG nova.compute.manager [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1132.790043] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.790935] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077e2c47-7ceb-4809-b5d3-478bbd247528 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.800711] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.800984] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14bc91a0-7770-4b3b-8a08-1c40347881c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.809687] env[65385]: DEBUG oslo_vmware.api [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1132.809687] env[65385]: value = "task-4454496" [ 1132.809687] env[65385]: _type = "Task" [ 1132.809687] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.824386] env[65385]: DEBUG oslo_vmware.api [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454496, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.836725] env[65385]: DEBUG nova.compute.manager [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1132.837591] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b06fff7-67f6-4869-ac68-09e7648d8a8a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.908240] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.908463] env[65385]: DEBUG nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1132.932312] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52777dcd-0196-c651-ab51-a64cd5dbff88, 'name': SearchDatastore_Task, 'duration_secs': 0.010551} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.932989] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.932989] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5873d9d9-ff2d-4228-a4ed-b07717ba082e/5873d9d9-ff2d-4228-a4ed-b07717ba082e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1132.933204] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bfbcbd5-d4c0-4374-8045-21e52e0bcd58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.941320] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1132.941320] env[65385]: value = "task-4454497" [ 1132.941320] env[65385]: _type = "Task" [ 1132.941320] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.954762] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454497, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.966228] env[65385]: DEBUG oslo_concurrency.lockutils [None req-202549cb-e5f7-4f15-9c62-f91e865323d7 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "54faee8c-6b2d-4864-bf96-26b9dfba583b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.795s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.024359] env[65385]: DEBUG oslo_vmware.api [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454489, 'name': ReconfigVM_Task, 'duration_secs': 5.755914} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.024787] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.025156] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Reconfigured VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1133.025786] env[65385]: WARNING neutronclient.v2_0.client [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.026352] env[65385]: WARNING neutronclient.v2_0.client [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.027684] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.028911] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.060562] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 0b02fcd5-9fc7-4543-a754-d31fa0186981] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1133.086088] env[65385]: WARNING neutronclient.v2_0.client [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.147031] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1133.147444] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0234dcc8-8d78-4455-96c9-ac016bf7026a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.157771] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1133.157771] env[65385]: value = "task-4454498" [ 1133.157771] env[65385]: _type = "Task" [ 1133.157771] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.172070] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.323083] env[65385]: DEBUG oslo_vmware.api [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454496, 'name': PowerOffVM_Task, 'duration_secs': 0.246174} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.323392] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1133.323582] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1133.323866] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-592b9a38-7b25-4d15-ad5e-7f9a87d57b9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.414295] env[65385]: DEBUG nova.compute.utils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1133.418071] env[65385]: DEBUG nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1133.418071] env[65385]: DEBUG nova.network.neutron [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1133.422029] env[65385]: WARNING neutronclient.v2_0.client [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.422029] env[65385]: WARNING neutronclient.v2_0.client [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.422029] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.422029] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.452772] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454497, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.505366] env[65385]: DEBUG nova.policy [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac74a729b62042eda193030d716719e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1598307c304d46a884a88b3eb5f5d7dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1133.526522] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1133.526881] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1133.526936] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleting the datastore file [datastore1] 7f22736b-654f-4d51-82d4-3fe80c1d0b52 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.527248] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ceb6f467-30f1-414b-a614-810c1b66cf28 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.533918] env[65385]: DEBUG oslo_vmware.api [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1133.533918] env[65385]: value = "task-4454500" [ 1133.533918] env[65385]: _type = "Task" [ 1133.533918] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.545512] env[65385]: DEBUG oslo_vmware.api [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454500, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.564206] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 370b6aa8-2337-48e1-8102-611ce4494a55] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1133.669846] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454498, 'name': PowerOffVM_Task, 'duration_secs': 0.233696} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.670132] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1133.670315] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance '6c826914-4c6b-456b-a403-2373d30c803f' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1133.846635] env[65385]: DEBUG nova.network.neutron [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Successfully created port: cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1133.853725] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1133.854373] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3584b78-4cd4-451f-9d77-b47d6803fb1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.863082] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1133.863082] env[65385]: value = "task-4454501" [ 1133.863082] env[65385]: _type = "Task" [ 1133.863082] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.873111] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.931625] env[65385]: DEBUG nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1133.953376] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454497, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523008} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.955674] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 5873d9d9-ff2d-4228-a4ed-b07717ba082e/5873d9d9-ff2d-4228-a4ed-b07717ba082e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1133.955674] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1133.955674] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a42b060b-4095-4633-82b2-4b7937c53b29 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.965402] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1133.965402] env[65385]: value = "task-4454502" [ 1133.965402] env[65385]: _type = "Task" [ 1133.965402] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.977265] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.044626] env[65385]: DEBUG oslo_vmware.api [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454500, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167894} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.045017] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.045276] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1134.045474] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1134.045652] env[65385]: INFO nova.compute.manager [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Took 1.26 seconds to destroy the instance on the hypervisor. [ 1134.045953] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1134.046191] env[65385]: DEBUG nova.compute.manager [-] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1134.046306] env[65385]: DEBUG nova.network.neutron [-] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1134.046576] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.047234] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.047512] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.068425] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 9908d395-545d-4caf-9757-320d6253d61e] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1134.086492] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.177637] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1134.177870] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1134.178050] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1134.178292] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1134.178415] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1134.178580] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1134.178849] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1134.178994] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1134.179286] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1134.179466] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1134.179636] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1134.184950] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7758f337-9989-4f32-aeb9-ff48defb848d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.204481] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1134.204481] env[65385]: value = "task-4454503" [ 1134.204481] env[65385]: _type = "Task" [ 1134.204481] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.215677] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454503, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.375504] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454501, 'name': PowerOffVM_Task, 'duration_secs': 0.19553} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.377062] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1134.377393] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1134.378724] env[65385]: DEBUG nova.compute.manager [req-56173e1e-7673-4606-94ec-94ab7ee4f05b req-52f8da03-b9d5-461c-ac88-41efe6ca9a21 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Received event network-vif-deleted-e4c07525-65a3-431c-a14b-f434713c739f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1134.378967] env[65385]: INFO nova.compute.manager [req-56173e1e-7673-4606-94ec-94ab7ee4f05b req-52f8da03-b9d5-461c-ac88-41efe6ca9a21 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Neutron deleted interface e4c07525-65a3-431c-a14b-f434713c739f; detaching it from the instance and deleting it from the info cache [ 1134.379191] env[65385]: DEBUG nova.network.neutron [req-56173e1e-7673-4606-94ec-94ab7ee4f05b req-52f8da03-b9d5-461c-ac88-41efe6ca9a21 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.381937] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6674df6d-cb42-4280-8854-c1b743d7f7bf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.391631] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1134.392591] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9819316-135e-4504-a276-541beaa6fddc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.423530] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1134.423994] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1134.424220] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Deleting the datastore file [datastore2] 54faee8c-6b2d-4864-bf96-26b9dfba583b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1134.424505] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fa6966c-8e88-448d-9099-21c0ac2fea3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.432662] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1134.432662] env[65385]: value = "task-4454505" [ 1134.432662] env[65385]: _type = "Task" [ 1134.432662] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.449098] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454505, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.478844] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.479049] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.479230] env[65385]: DEBUG nova.network.neutron [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1134.480587] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072348} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.480833] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.482467] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4679e61c-d940-45dd-a9a2-df91af7dee60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.522326] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 5873d9d9-ff2d-4228-a4ed-b07717ba082e/5873d9d9-ff2d-4228-a4ed-b07717ba082e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.523350] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-978b58d1-2cde-4122-bc78-812e7ff2b25b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.546106] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1134.546106] env[65385]: value = "task-4454506" [ 1134.546106] env[65385]: _type = "Task" [ 1134.546106] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.560541] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454506, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.572046] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 0c6206ee-c787-4cd1-9289-1b2620dc4c5f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1134.715467] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454503, 'name': ReconfigVM_Task, 'duration_secs': 0.320098} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.715768] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance '6c826914-4c6b-456b-a403-2373d30c803f' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1134.781034] env[65385]: DEBUG nova.network.neutron [-] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.886839] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6ea05c7-28d8-49a4-9c19-1e00717abf7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.898828] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91e2217-8c04-4fd8-a10a-4f2bc191425a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.934647] env[65385]: DEBUG nova.compute.manager [req-56173e1e-7673-4606-94ec-94ab7ee4f05b req-52f8da03-b9d5-461c-ac88-41efe6ca9a21 service nova] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Detach interface failed, port_id=e4c07525-65a3-431c-a14b-f434713c739f, reason: Instance 7f22736b-654f-4d51-82d4-3fe80c1d0b52 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1134.943669] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454505, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369425} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.944963] env[65385]: DEBUG nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1134.947041] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.947242] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1134.947417] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1134.975216] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1134.975549] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1134.975707] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1134.975895] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1134.976040] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1134.976190] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1134.976396] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1134.976570] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1134.976704] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1134.976864] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1134.977038] env[65385]: DEBUG nova.virt.hardware [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1134.977900] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544191de-8b39-4bc9-bae5-b2e284c8cf38 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.986571] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6a9d2f-c06b-4e03-84d4-60d2d7b0786d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.991125] env[65385]: WARNING neutronclient.v2_0.client [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.991738] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.992094] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.059330] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454506, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.075466] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 703ed9b6-7cd8-4a84-9847-d34fb1c51a3f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1135.223124] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1135.223452] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1135.223831] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1135.223831] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1135.223999] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1135.224551] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1135.224835] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1135.225024] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1135.225222] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1135.225491] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1135.225618] env[65385]: DEBUG nova.virt.hardware [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1135.231077] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1135.231444] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68877eda-e5fe-41ad-b849-3df8ed921416 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.252054] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1135.252054] env[65385]: value = "task-4454507" [ 1135.252054] env[65385]: _type = "Task" [ 1135.252054] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.261967] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454507, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.283402] env[65385]: INFO nova.compute.manager [-] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Took 1.24 seconds to deallocate network for instance. [ 1135.340922] env[65385]: DEBUG nova.compute.manager [req-2e32ab21-6c06-49ee-89c5-938d13920672 req-42c5532a-b0db-485d-9d95-8eabd9a30dd0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-vif-plugged-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1135.341296] env[65385]: DEBUG oslo_concurrency.lockutils [req-2e32ab21-6c06-49ee-89c5-938d13920672 req-42c5532a-b0db-485d-9d95-8eabd9a30dd0 service nova] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.342203] env[65385]: DEBUG oslo_concurrency.lockutils [req-2e32ab21-6c06-49ee-89c5-938d13920672 req-42c5532a-b0db-485d-9d95-8eabd9a30dd0 service nova] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.342203] env[65385]: DEBUG oslo_concurrency.lockutils [req-2e32ab21-6c06-49ee-89c5-938d13920672 req-42c5532a-b0db-485d-9d95-8eabd9a30dd0 service nova] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.342203] env[65385]: DEBUG nova.compute.manager [req-2e32ab21-6c06-49ee-89c5-938d13920672 req-42c5532a-b0db-485d-9d95-8eabd9a30dd0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] No waiting events found dispatching network-vif-plugged-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1135.342203] env[65385]: WARNING nova.compute.manager [req-2e32ab21-6c06-49ee-89c5-938d13920672 req-42c5532a-b0db-485d-9d95-8eabd9a30dd0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received unexpected event network-vif-plugged-cf3df048-713a-474d-9278-a467e50ec4f8 for instance with vm_state building and task_state spawning. [ 1135.426597] env[65385]: DEBUG nova.network.neutron [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Successfully updated port: cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1135.488603] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.489061] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.556193] env[65385]: WARNING neutronclient.v2_0.client [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.556944] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.557518] env[65385]: WARNING openstack [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.571593] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454506, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.578959] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: c18b5bde-1a17-4954-a2c5-582815ebfec8] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1135.711438] env[65385]: INFO nova.network.neutron [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Port 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1135.711845] env[65385]: DEBUG nova.network.neutron [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1135.762616] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454507, 'name': ReconfigVM_Task, 'duration_secs': 0.380123} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.763106] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1135.763901] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c974e192-dab6-4019-a33a-caece5f3c048 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.794881] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.795925] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.796246] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.796469] env[65385]: DEBUG nova.objects.instance [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'resources' on Instance uuid 7f22736b-654f-4d51-82d4-3fe80c1d0b52 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.797713] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34382d37-f053-4a10-ad74-d18e1fd240bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.820539] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1135.820539] env[65385]: value = "task-4454508" [ 1135.820539] env[65385]: _type = "Task" [ 1135.820539] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.831023] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454508, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.930196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.930402] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.930582] env[65385]: DEBUG nova.network.neutron [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1135.983338] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1135.983620] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1135.983787] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1135.983968] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1135.984145] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1135.984290] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1135.984491] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1135.984645] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1135.984802] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1135.984954] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1135.985134] env[65385]: DEBUG nova.virt.hardware [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1135.986000] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6b9598-2df9-4a71-9e8e-fa1bd203bad4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.994139] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a82a4c1-486b-495a-b394-79bacb252d9f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.008786] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1136.014397] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1136.014716] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1136.014966] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de66c021-37bb-42e5-a5ed-c1b09743dae0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.033637] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1136.033637] env[65385]: value = "task-4454509" [ 1136.033637] env[65385]: _type = "Task" [ 1136.033637] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.044735] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454509, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.057817] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454506, 'name': ReconfigVM_Task, 'duration_secs': 1.270993} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.058130] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 5873d9d9-ff2d-4228-a4ed-b07717ba082e/5873d9d9-ff2d-4228-a4ed-b07717ba082e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.058981] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-436411d7-7ec7-45b5-81c6-3ce75cd98bca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.067880] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1136.067880] env[65385]: value = "task-4454510" [ 1136.067880] env[65385]: _type = "Task" [ 1136.067880] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.077404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.077645] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.077991] env[65385]: DEBUG nova.objects.instance [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'flavor' on Instance uuid b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.079329] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454510, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.082039] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: af8d01e2-fb86-41c4-99a5-204b30eeda0c] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1136.219648] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.333108] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454508, 'name': ReconfigVM_Task, 'duration_secs': 0.412083} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.333521] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.333912] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance '6c826914-4c6b-456b-a403-2373d30c803f' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1136.416353] env[65385]: DEBUG nova.compute.manager [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1136.416353] env[65385]: DEBUG nova.compute.manager [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing instance network info cache due to event network-changed-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1136.416353] env[65385]: DEBUG oslo_concurrency.lockutils [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Acquiring lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.416353] env[65385]: DEBUG oslo_concurrency.lockutils [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Acquired lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.416353] env[65385]: DEBUG nova.network.neutron [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Refreshing network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1136.434016] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.434476] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.474336] env[65385]: DEBUG nova.network.neutron [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1136.500186] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.500186] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.520424] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f0831a-8527-4e2d-8b75-bf5cfd547bb5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.529665] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f19807-8def-40ae-917e-4d71e5178d8e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.567362] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d182bc-7165-4dc1-9266-15725323586e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.573706] env[65385]: WARNING neutronclient.v2_0.client [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.574474] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.574914] env[65385]: WARNING openstack [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.584469] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454509, 'name': CreateVM_Task, 'duration_secs': 0.395823} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.585889] env[65385]: WARNING neutronclient.v2_0.client [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.586515] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.586877] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.600220] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1136.600515] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 0c345794-036c-45d1-985e-cf5c116bd873] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1136.605269] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.605473] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.605778] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1136.606740] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3d6f951-9f15-4f05-b23c-124b2b892ec9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.609974] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61ee472-8787-470d-9b19-4af5ddeb4978 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.618015] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454510, 'name': Rename_Task, 'duration_secs': 0.166408} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.618983] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1136.619293] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a966561-92a7-4f54-8c14-50ee4ab1d2e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.622938] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1136.622938] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b24987-36cf-9c28-2f27-af5488799915" [ 1136.622938] env[65385]: _type = "Task" [ 1136.622938] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.631174] env[65385]: DEBUG nova.compute.provider_tree [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.640916] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1136.640916] env[65385]: value = "task-4454511" [ 1136.640916] env[65385]: _type = "Task" [ 1136.640916] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.653136] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b24987-36cf-9c28-2f27-af5488799915, 'name': SearchDatastore_Task, 'duration_secs': 0.010695} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.653913] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.653913] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1136.654209] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.654757] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.654757] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1136.655204] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8d96007-e88d-4340-b3b1-b02b834f946a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.660960] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454511, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.669115] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1136.669115] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1136.670923] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-932e4381-00fc-452a-86f8-18cdf54554b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.676655] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1136.676655] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b4bdec-4162-6d0a-d46c-1c62b47ef7d7" [ 1136.676655] env[65385]: _type = "Task" [ 1136.676655] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.687262] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b4bdec-4162-6d0a-d46c-1c62b47ef7d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.689061] env[65385]: DEBUG nova.network.neutron [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf3df048-71", "ovs_interfaceid": "cf3df048-713a-474d-9278-a467e50ec4f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.724613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1a918d03-6620-4d36-9464-3990400c13c5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-900639b6-9b98-436f-aaad-ea8391f67393-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.347s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.752846] env[65385]: DEBUG nova.objects.instance [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'pci_requests' on Instance uuid b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.841882] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99aa6b7b-417d-40fe-aff7-072ed9446535 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.867598] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916df81a-40d9-4aec-af72-a76345446204 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.892320] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance '6c826914-4c6b-456b-a403-2373d30c803f' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1136.920528] env[65385]: WARNING neutronclient.v2_0.client [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.921383] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.921383] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.031133] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.031540] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.101332] env[65385]: WARNING neutronclient.v2_0.client [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.102052] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.102482] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.110585] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: f25cc41e-560e-4cab-8bf0-dac47899d037] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1137.145610] env[65385]: DEBUG nova.scheduler.client.report [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1137.164169] env[65385]: DEBUG oslo_vmware.api [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454511, 'name': PowerOnVM_Task, 'duration_secs': 0.534999} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.165164] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1137.165387] env[65385]: INFO nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Took 8.13 seconds to spawn the instance on the hypervisor. [ 1137.165538] env[65385]: DEBUG nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1137.166361] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905a9b92-dd22-49f7-bf9d-1a47e2274b02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.187578] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b4bdec-4162-6d0a-d46c-1c62b47ef7d7, 'name': SearchDatastore_Task, 'duration_secs': 0.009976} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.188304] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc745831-ec07-4d94-a7f6-bcdba9981bd1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.191443] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.191668] env[65385]: DEBUG nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Instance network_info: |[{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf3df048-71", "ovs_interfaceid": "cf3df048-713a-474d-9278-a467e50ec4f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1137.192142] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:47:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9875d38f-76e2-416c-bfb7-f18a22b0d8ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf3df048-713a-474d-9278-a467e50ec4f8', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1137.200613] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1137.202958] env[65385]: DEBUG nova.network.neutron [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updated VIF entry in instance network info cache for port 2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1137.203311] env[65385]: DEBUG nova.network.neutron [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [{"id": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "address": "fa:16:3e:ac:b2:c8", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ab1f4d8-5e", "ovs_interfaceid": "2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1137.206296] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1137.206475] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1137.206475] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525b6e16-abdd-ad01-c33e-d38ce965996a" [ 1137.206475] env[65385]: _type = "Task" [ 1137.206475] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.206856] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4b614fb-ac05-43c9-bada-942df547e1c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.232668] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525b6e16-abdd-ad01-c33e-d38ce965996a, 'name': SearchDatastore_Task, 'duration_secs': 0.011947} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.234013] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.234267] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1137.234503] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1137.234503] env[65385]: value = "task-4454512" [ 1137.234503] env[65385]: _type = "Task" [ 1137.234503] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.234800] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-818a65be-3104-427f-aba4-bfde283d507d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.246403] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454512, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.247852] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1137.247852] env[65385]: value = "task-4454513" [ 1137.247852] env[65385]: _type = "Task" [ 1137.247852] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.256742] env[65385]: DEBUG nova.objects.base [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1137.256923] env[65385]: DEBUG nova.network.neutron [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1137.257371] env[65385]: WARNING neutronclient.v2_0.client [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.257649] env[65385]: WARNING neutronclient.v2_0.client [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.258247] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.258653] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.266327] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454513, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.340925] env[65385]: DEBUG nova.policy [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699e2908dc5b4c889a69e5955194537f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1991528775bd4557ae6150b30095e923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1137.398471] env[65385]: WARNING neutronclient.v2_0.client [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.398818] env[65385]: WARNING neutronclient.v2_0.client [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.462351] env[65385]: DEBUG nova.network.neutron [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Port 92c45f41-92b4-40c3-8516-30189d29e33d binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1137.531059] env[65385]: DEBUG nova.compute.manager [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1137.531298] env[65385]: DEBUG nova.compute.manager [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing instance network info cache due to event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1137.531634] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Acquiring lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.531777] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Acquired lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1137.531973] env[65385]: DEBUG nova.network.neutron [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1137.613794] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: ad29afd9-3e52-4fac-977f-91774acad7d0] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1137.651586] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1137.684347] env[65385]: INFO nova.scheduler.client.report [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted allocations for instance 7f22736b-654f-4d51-82d4-3fe80c1d0b52 [ 1137.690443] env[65385]: INFO nova.compute.manager [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Took 12.12 seconds to build instance. [ 1137.708194] env[65385]: DEBUG oslo_concurrency.lockutils [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Releasing lock "refresh_cache-900639b6-9b98-436f-aaad-ea8391f67393" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.708476] env[65385]: DEBUG nova.compute.manager [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1137.708620] env[65385]: DEBUG nova.compute.manager [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing instance network info cache due to event network-changed-6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1137.708844] env[65385]: DEBUG oslo_concurrency.lockutils [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.709024] env[65385]: DEBUG oslo_concurrency.lockutils [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1137.709226] env[65385]: DEBUG nova.network.neutron [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1137.749057] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454512, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.758738] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454513, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.013699] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.014024] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.014244] env[65385]: INFO nova.compute.manager [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Shelving [ 1138.034421] env[65385]: WARNING neutronclient.v2_0.client [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.035182] env[65385]: WARNING openstack [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.035543] env[65385]: WARNING openstack [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.117174] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 26a2d464-0aab-463f-8265-948a4f0ff188] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1138.177482] env[65385]: WARNING openstack [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.177999] env[65385]: WARNING openstack [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.195226] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e247cd45-1a8f-4038-b836-f5a3703e7bef tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.634s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.195747] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e836f4a0-21fa-4389-9659-337df8c79697 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "7f22736b-654f-4d51-82d4-3fe80c1d0b52" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.918s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.212328] env[65385]: WARNING neutronclient.v2_0.client [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.213075] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.213490] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.248870] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454512, 'name': CreateVM_Task, 'duration_secs': 0.582086} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.252658] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1138.253750] env[65385]: WARNING neutronclient.v2_0.client [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.254267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.254445] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1138.254871] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1138.255491] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b964d19-d1cf-4195-8130-7ece5456daeb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.261442] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454513, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557645} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.263493] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1138.263889] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1138.264380] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1138.264380] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529545d4-e30a-e9b4-33ee-f81bc7705269" [ 1138.264380] env[65385]: _type = "Task" [ 1138.264380] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.264677] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd0a4c53-deaa-40df-bff2-8c7693327a5b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.268360] env[65385]: WARNING neutronclient.v2_0.client [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.269164] env[65385]: WARNING openstack [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.269621] env[65385]: WARNING openstack [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.287947] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529545d4-e30a-e9b4-33ee-f81bc7705269, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.289390] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1138.289390] env[65385]: value = "task-4454514" [ 1138.289390] env[65385]: _type = "Task" [ 1138.289390] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.299759] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454514, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.441711] env[65385]: DEBUG nova.network.neutron [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updated VIF entry in instance network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1138.442405] env[65385]: DEBUG nova.network.neutron [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf3df048-71", "ovs_interfaceid": "cf3df048-713a-474d-9278-a467e50ec4f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1138.473687] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.474104] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.491952] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "6c826914-4c6b-456b-a403-2373d30c803f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.492205] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.492377] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.618924] env[65385]: WARNING neutronclient.v2_0.client [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.619636] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.619988] env[65385]: WARNING openstack [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.628078] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 65f9bc55-7bab-43b0-a974-eb5080389b7d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1138.764854] env[65385]: DEBUG nova.network.neutron [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updated VIF entry in instance network info cache for port 6e6c199b-75f0-473b-90b5-0daaf633ba14. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1138.767820] env[65385]: DEBUG nova.network.neutron [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1138.782431] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]529545d4-e30a-e9b4-33ee-f81bc7705269, 'name': SearchDatastore_Task, 'duration_secs': 0.038182} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.782431] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1138.782431] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1138.782431] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.782431] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1138.782431] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1138.782675] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84cb7005-fd36-4a01-b7c3-9bda6b2275c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.795275] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1138.795513] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1138.796606] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02d152d6-308f-4f37-a190-3bd48c706b82 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.801928] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454514, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.266577} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.802545] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1138.805836] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdabb27-cd15-49ab-a5e9-239c311358ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.807520] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1138.807520] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f77e26-9269-4701-d888-f06fd3ffdb70" [ 1138.807520] env[65385]: _type = "Task" [ 1138.807520] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.828400] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1138.828400] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7aae442-8726-4209-9599-b1c74d41befb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.846578] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f77e26-9269-4701-d888-f06fd3ffdb70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.854648] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1138.854648] env[65385]: value = "task-4454515" [ 1138.854648] env[65385]: _type = "Task" [ 1138.854648] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.863657] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454515, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.897747] env[65385]: DEBUG nova.compute.manager [req-2ec2fcf2-a264-4167-afca-6eeb313ee57d req-b5ec8792-9a96-4c01-99c8-98423ab63266 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-vif-plugged-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1138.898079] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ec2fcf2-a264-4167-afca-6eeb313ee57d req-b5ec8792-9a96-4c01-99c8-98423ab63266 service nova] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.898167] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ec2fcf2-a264-4167-afca-6eeb313ee57d req-b5ec8792-9a96-4c01-99c8-98423ab63266 service nova] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.898330] env[65385]: DEBUG oslo_concurrency.lockutils [req-2ec2fcf2-a264-4167-afca-6eeb313ee57d req-b5ec8792-9a96-4c01-99c8-98423ab63266 service nova] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.898486] env[65385]: DEBUG nova.compute.manager [req-2ec2fcf2-a264-4167-afca-6eeb313ee57d req-b5ec8792-9a96-4c01-99c8-98423ab63266 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] No waiting events found dispatching network-vif-plugged-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1138.898644] env[65385]: WARNING nova.compute.manager [req-2ec2fcf2-a264-4167-afca-6eeb313ee57d req-b5ec8792-9a96-4c01-99c8-98423ab63266 service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received unexpected event network-vif-plugged-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 for instance with vm_state active and task_state None. [ 1138.945595] env[65385]: DEBUG oslo_concurrency.lockutils [req-2a99eb23-7bc5-42a4-b9ae-0e0045c1b083 req-647335b6-0423-428d-a1ea-2b0a778475e0 service nova] Releasing lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1138.992606] env[65385]: DEBUG nova.network.neutron [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Successfully updated port: 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1139.024623] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1139.024957] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0f4b747-2682-4bc7-b115-99be528776c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.032957] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1139.032957] env[65385]: value = "task-4454516" [ 1139.032957] env[65385]: _type = "Task" [ 1139.032957] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.042941] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.131653] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: ab8df1d8-615d-4083-bdf4-25f679a64982] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1139.274148] env[65385]: DEBUG oslo_concurrency.lockutils [req-5a1ea6f6-4a61-45e1-b434-7fee9f291b3d req-45e8ae3e-31f6-4637-b65a-a7c5ee3f78ab service nova] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1139.322184] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f77e26-9269-4701-d888-f06fd3ffdb70, 'name': SearchDatastore_Task, 'duration_secs': 0.039352} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.322869] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91efda25-84ba-4e80-af62-7aa940ece206 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.329605] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1139.329605] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52da18ae-25da-3b45-4e45-d77b037e1ecc" [ 1139.329605] env[65385]: _type = "Task" [ 1139.329605] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.338478] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52da18ae-25da-3b45-4e45-d77b037e1ecc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.364983] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454515, 'name': ReconfigVM_Task, 'duration_secs': 0.369947} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.365423] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 54faee8c-6b2d-4864-bf96-26b9dfba583b/54faee8c-6b2d-4864-bf96-26b9dfba583b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1139.366357] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7bd2c33f-6fdd-4071-a568-6de1c40fb155 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.373225] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1139.373225] env[65385]: value = "task-4454517" [ 1139.373225] env[65385]: _type = "Task" [ 1139.373225] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.388592] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454517, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.496169] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.496169] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1139.496169] env[65385]: DEBUG nova.network.neutron [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1139.498741] env[65385]: WARNING neutronclient.v2_0.client [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.532838] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.533580] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1139.533580] env[65385]: DEBUG nova.network.neutron [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1139.545125] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454516, 'name': PowerOffVM_Task, 'duration_secs': 0.217619} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.545443] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1139.546332] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c93875e-3c89-457b-b27e-74d9350e4cd4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.569114] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7e6323-e90a-4a91-a2f5-31abc60958d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.635757] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: e9fd0c9f-5795-44df-9dcb-982b7722ae69] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1139.670678] env[65385]: DEBUG nova.compute.manager [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1139.841325] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52da18ae-25da-3b45-4e45-d77b037e1ecc, 'name': SearchDatastore_Task, 'duration_secs': 0.010869} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.841629] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1139.841888] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1139.842201] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd6097f3-ba65-4d62-b7a1-777ea6817698 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.851571] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1139.851571] env[65385]: value = "task-4454518" [ 1139.851571] env[65385]: _type = "Task" [ 1139.851571] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.861648] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454518, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.885981] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454517, 'name': Rename_Task, 'duration_secs': 0.147975} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.886320] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1139.886602] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-970f7d94-f5c6-4fbc-90c7-1b4c7a827400 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.895419] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1139.895419] env[65385]: value = "task-4454519" [ 1139.895419] env[65385]: _type = "Task" [ 1139.895419] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.906499] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454519, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.913120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "8db87973-e093-4da1-8534-62ea6896a57d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.913368] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "8db87973-e093-4da1-8534-62ea6896a57d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.003255] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.003973] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.039367] env[65385]: WARNING neutronclient.v2_0.client [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.040097] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.040452] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.054183] env[65385]: WARNING nova.network.neutron [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] 16967e55-c149-4a58-905f-c47b7facac11 already exists in list: networks containing: ['16967e55-c149-4a58-905f-c47b7facac11']. ignoring it [ 1140.084044] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1140.084379] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fb5e2732-208a-40e6-a4e2-72a314d2c035 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.093495] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1140.093495] env[65385]: value = "task-4454520" [ 1140.093495] env[65385]: _type = "Task" [ 1140.093495] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.098745] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.099422] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.113698] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454520, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.138784] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: c5d00bb5-d869-466f-826e-61622e57959f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1140.167851] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.168330] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.180058] env[65385]: WARNING neutronclient.v2_0.client [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.180977] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.181515] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.194063] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.194352] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.256192] env[65385]: WARNING neutronclient.v2_0.client [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.256903] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.257266] env[65385]: WARNING openstack [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.300129] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.300275] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.358601] env[65385]: DEBUG nova.network.neutron [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1140.368194] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454518, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.379370] env[65385]: WARNING neutronclient.v2_0.client [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.380046] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.380424] env[65385]: WARNING openstack [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.408433] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454519, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.416326] env[65385]: DEBUG nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1140.475964] env[65385]: DEBUG nova.network.neutron [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "address": "fa:16:3e:6b:e5:fd", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap330c3a6f-16", "ovs_interfaceid": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1140.604324] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454520, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.642715] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 98430d7e-626e-4532-802e-38f7fc583ae3] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1140.699460] env[65385]: INFO nova.compute.claims [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.862813] env[65385]: DEBUG oslo_concurrency.lockutils [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1140.869993] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454518, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54289} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.870471] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1140.870691] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1140.871094] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cf7c8a3-d92d-4736-816e-b21a82fe6997 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.882240] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1140.882240] env[65385]: value = "task-4454521" [ 1140.882240] env[65385]: _type = "Task" [ 1140.882240] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.892751] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454521, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.906299] env[65385]: DEBUG oslo_vmware.api [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454519, 'name': PowerOnVM_Task, 'duration_secs': 0.988368} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.906299] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1140.906299] env[65385]: DEBUG nova.compute.manager [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1140.906714] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566a4884-1715-431b-9166-580785acf8aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.941399] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.950732] env[65385]: DEBUG nova.compute.manager [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-changed-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1140.950932] env[65385]: DEBUG nova.compute.manager [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing instance network info cache due to event network-changed-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1140.951128] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.979078] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1140.979782] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.979939] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1140.980241] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1140.980420] env[65385]: DEBUG nova.network.neutron [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Refreshing network info cache for port 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1140.982312] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11535ca-db8b-40ba-8611-8204ca6e9a88 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.002034] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1141.002034] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1141.002034] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1141.002034] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1141.002318] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1141.002318] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1141.002434] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1141.002565] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1141.002692] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1141.002847] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1141.003029] env[65385]: DEBUG nova.virt.hardware [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1141.009367] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Reconfiguring VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1141.009985] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b6dc118-3acd-44c6-aeb7-946083133f17 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.029946] env[65385]: DEBUG oslo_vmware.api [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1141.029946] env[65385]: value = "task-4454522" [ 1141.029946] env[65385]: _type = "Task" [ 1141.029946] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.040578] env[65385]: DEBUG oslo_vmware.api [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454522, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.106537] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454520, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.146268] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: f521e37d-aa86-4533-9fbe-7d040ed90e53] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1141.206320] env[65385]: INFO nova.compute.resource_tracker [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating resource usage from migration 63618a0e-63c2-49ca-84d5-466f3f3f0c0d [ 1141.388225] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340a7e8c-11b4-4db2-9c4f-76b294df7851 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.392232] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb8228c-da04-4ad8-b600-59f3dbf7b90d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.397862] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454521, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07478} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.398471] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1141.399299] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35279365-af52-451e-a2b3-72aa01182e58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.416499] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e152cfa2-9fbd-4b76-9e38-9cc89ccedad9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.419619] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3283c50-241c-4da2-85e7-43f5c5546902 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.444497] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1141.447032] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3df29aa-fb39-432d-8c63-22200e47e24a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.464883] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.488880] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance '6c826914-4c6b-456b-a403-2373d30c803f' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1141.495184] env[65385]: WARNING neutronclient.v2_0.client [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.495813] env[65385]: WARNING openstack [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.496165] env[65385]: WARNING openstack [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.504290] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc63ce4-696e-4604-a083-8bc5a4facb8d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.509363] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1141.509363] env[65385]: value = "task-4454523" [ 1141.509363] env[65385]: _type = "Task" [ 1141.509363] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.516076] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d0b25a-3dae-4921-877f-7a29bae70b18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.529477] env[65385]: DEBUG nova.compute.provider_tree [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.535027] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454523, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.543937] env[65385]: DEBUG oslo_vmware.api [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.604285] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454520, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.605688] env[65385]: WARNING openstack [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.606053] env[65385]: WARNING openstack [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.650691] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: b702ba27-bbc7-4030-a96d-30e1ed656a9b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1141.669516] env[65385]: WARNING neutronclient.v2_0.client [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.670258] env[65385]: WARNING openstack [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.670649] env[65385]: WARNING openstack [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.762881] env[65385]: DEBUG nova.network.neutron [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updated VIF entry in instance network info cache for port 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1141.763402] env[65385]: DEBUG nova.network.neutron [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "address": "fa:16:3e:6b:e5:fd", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap330c3a6f-16", "ovs_interfaceid": "330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1142.005717] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1142.006122] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ff70c96-1d4d-453a-97aa-f87b1aa74b77 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.017523] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1142.017523] env[65385]: value = "task-4454524" [ 1142.017523] env[65385]: _type = "Task" [ 1142.017523] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.023856] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454523, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.029079] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454524, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.035275] env[65385]: DEBUG nova.scheduler.client.report [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.049527] env[65385]: DEBUG oslo_vmware.api [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.106353] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454520, 'name': CreateSnapshot_Task, 'duration_secs': 1.598713} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.106627] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1142.107395] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17f60ea-2ad1-4d46-9775-3861397beddb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.153234] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 1ad1e02b-fb82-4a88-9047-be316d590bd2] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1142.266155] env[65385]: DEBUG oslo_concurrency.lockutils [req-ddbaef74-6c81-4f54-89a3-f2715debd2e9 req-d20bf506-dcf8-40fd-abe1-f06851ff3d4a service nova] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.323760] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "54faee8c-6b2d-4864-bf96-26b9dfba583b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.324104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "54faee8c-6b2d-4864-bf96-26b9dfba583b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.324342] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "54faee8c-6b2d-4864-bf96-26b9dfba583b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.324527] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "54faee8c-6b2d-4864-bf96-26b9dfba583b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.324709] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "54faee8c-6b2d-4864-bf96-26b9dfba583b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.326824] env[65385]: INFO nova.compute.manager [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Terminating instance [ 1142.526217] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454523, 'name': ReconfigVM_Task, 'duration_secs': 0.702198} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.527128] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1142.527853] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48e0a82a-5942-4ae9-be1c-cd0180e9416e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.535111] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454524, 'name': PowerOnVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.540260] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1142.540260] env[65385]: value = "task-4454525" [ 1142.540260] env[65385]: _type = "Task" [ 1142.540260] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.548660] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.354s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.548886] env[65385]: INFO nova.compute.manager [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Migrating [ 1142.555908] env[65385]: DEBUG oslo_vmware.api [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454522, 'name': ReconfigVM_Task, 'duration_secs': 1.516153} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.559122] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.618s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.561695] env[65385]: INFO nova.compute.claims [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.566411] env[65385]: WARNING neutronclient.v2_0.client [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1142.566831] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.567159] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Reconfigured VM to attach interface {{(pid=65385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1142.584286] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454525, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.629298] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1142.630581] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-195f729c-2159-4864-b381-0aa99ce10c4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.641146] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1142.641146] env[65385]: value = "task-4454526" [ 1142.641146] env[65385]: _type = "Task" [ 1142.641146] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.653132] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454526, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.655987] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 6746e11f-e9fa-4d3b-a8cf-fa64ebcee4e0] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1142.830296] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "refresh_cache-54faee8c-6b2d-4864-bf96-26b9dfba583b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.830512] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquired lock "refresh_cache-54faee8c-6b2d-4864-bf96-26b9dfba583b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.830693] env[65385]: DEBUG nova.network.neutron [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1143.033120] env[65385]: DEBUG oslo_vmware.api [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454524, 'name': PowerOnVM_Task, 'duration_secs': 0.617189} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.033686] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1143.034277] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-668bb138-423f-4d78-9b3a-ba98e87de056 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance '6c826914-4c6b-456b-a403-2373d30c803f' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1143.059513] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454525, 'name': Rename_Task, 'duration_secs': 0.182215} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.060077] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1143.060328] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e7e910a-1878-4f27-afc4-ceed142de582 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.069964] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1143.069964] env[65385]: value = "task-4454527" [ 1143.069964] env[65385]: _type = "Task" [ 1143.069964] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.086874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bdaa52c0-afbb-414f-b5b7-82eaab7f8673 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.009s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.088665] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.091575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.091575] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.091575] env[65385]: DEBUG nova.network.neutron [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1143.152691] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454526, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.159414] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 38e0ebf5-1d92-4906-b0bd-b0ebf8651f22] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1143.334682] env[65385]: WARNING neutronclient.v2_0.client [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.336068] env[65385]: WARNING openstack [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.336068] env[65385]: WARNING openstack [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.361533] env[65385]: DEBUG nova.network.neutron [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1143.419710] env[65385]: DEBUG nova.network.neutron [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1143.585041] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454527, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.594960] env[65385]: WARNING neutronclient.v2_0.client [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.595958] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.596337] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.655353] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454526, 'name': CloneVM_Task} progress is 95%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.663040] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 7fcbdc79-688c-479f-94e2-f4542abe714b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1143.743481] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.744027] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.818463] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be94c19d-55ce-414e-97ba-527554885d1f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.822302] env[65385]: WARNING neutronclient.v2_0.client [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.822959] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.823327] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.838040] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8795f00-8c7d-4e15-8b53-6574a9ce90ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.877689] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa3189b-ab5f-44c3-aae4-640a05727070 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.887414] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d54e2a3-fec5-41d3-96b1-e8357fb13504 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.903186] env[65385]: DEBUG nova.compute.provider_tree [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.922184] env[65385]: DEBUG nova.network.neutron [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance_info_cache with network_info: [{"id": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "address": "fa:16:3e:b0:35:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e4dcc22-4c", "ovs_interfaceid": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1143.923076] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Releasing lock "refresh_cache-54faee8c-6b2d-4864-bf96-26b9dfba583b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.923538] env[65385]: DEBUG nova.compute.manager [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1143.923754] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1143.924803] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44afd00-8186-4e94-ab93-345f574b0d66 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.933253] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.934077] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8cc7fac-970c-4e78-9da1-10f69b3fbbd7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.941802] env[65385]: DEBUG oslo_vmware.api [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1143.941802] env[65385]: value = "task-4454528" [ 1143.941802] env[65385]: _type = "Task" [ 1143.941802] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.951895] env[65385]: DEBUG oslo_vmware.api [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.085921] env[65385]: DEBUG oslo_vmware.api [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454527, 'name': PowerOnVM_Task, 'duration_secs': 0.800082} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.086236] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1144.086437] env[65385]: INFO nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Took 9.14 seconds to spawn the instance on the hypervisor. [ 1144.086709] env[65385]: DEBUG nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1144.087561] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106d2031-931b-408f-9336-cff59d05a670 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.152913] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454526, 'name': CloneVM_Task, 'duration_secs': 1.457452} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.153263] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Created linked-clone VM from snapshot [ 1144.154081] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4739af6a-502e-416e-b3c3-e3b2d0233ba4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.162805] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Uploading image 1b100802-b12e-4f1f-a998-6080dbb6821d {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1144.166573] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 483cf486-f0fc-48a4-9db2-970d0f35865a] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1144.192839] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1144.192839] env[65385]: value = "vm-871188" [ 1144.192839] env[65385]: _type = "VirtualMachine" [ 1144.192839] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1144.193142] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-48fd40fd-389c-40ca-9c34-4e1188b7807b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.202488] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease: (returnval){ [ 1144.202488] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5216f3bc-0791-8137-63c1-09abd3302de4" [ 1144.202488] env[65385]: _type = "HttpNfcLease" [ 1144.202488] env[65385]: } obtained for exporting VM: (result){ [ 1144.202488] env[65385]: value = "vm-871188" [ 1144.202488] env[65385]: _type = "VirtualMachine" [ 1144.202488] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1144.202767] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the lease: (returnval){ [ 1144.202767] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5216f3bc-0791-8137-63c1-09abd3302de4" [ 1144.202767] env[65385]: _type = "HttpNfcLease" [ 1144.202767] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1144.210351] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1144.210351] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5216f3bc-0791-8137-63c1-09abd3302de4" [ 1144.210351] env[65385]: _type = "HttpNfcLease" [ 1144.210351] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1144.407289] env[65385]: DEBUG nova.scheduler.client.report [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1144.424978] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.452180] env[65385]: DEBUG oslo_vmware.api [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454528, 'name': PowerOffVM_Task, 'duration_secs': 0.173075} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.452434] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1144.452594] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1144.452847] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8cdad1c7-44eb-4f3b-9f9b-a13b72518e12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.479443] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1144.479668] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1144.479843] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Deleting the datastore file [datastore1] 54faee8c-6b2d-4864-bf96-26b9dfba583b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1144.480122] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e04300a-9512-4a27-b27b-fa5017422538 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.486699] env[65385]: DEBUG oslo_vmware.api [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for the task: (returnval){ [ 1144.486699] env[65385]: value = "task-4454531" [ 1144.486699] env[65385]: _type = "Task" [ 1144.486699] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.495155] env[65385]: DEBUG oslo_vmware.api [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.606883] env[65385]: INFO nova.compute.manager [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Took 16.82 seconds to build instance. [ 1144.669848] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 90c36f67-0c25-4487-a5e4-89827290953b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1144.711126] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1144.711126] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5216f3bc-0791-8137-63c1-09abd3302de4" [ 1144.711126] env[65385]: _type = "HttpNfcLease" [ 1144.711126] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1144.711423] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1144.711423] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5216f3bc-0791-8137-63c1-09abd3302de4" [ 1144.711423] env[65385]: _type = "HttpNfcLease" [ 1144.711423] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1144.712166] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e595b2-fc9e-4348-86d0-031e4093ea72 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.720026] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3a040-c133-18d4-babf-590d1ec542ca/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1144.720241] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3a040-c133-18d4-babf-590d1ec542ca/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1144.788435] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "interface-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.788719] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.821596] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b0c3dfb8-ccf0-448f-b432-58e7ea24f3cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.912815] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.913390] env[65385]: DEBUG nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1144.916494] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.451s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.916694] env[65385]: DEBUG nova.objects.instance [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1145.001723] env[65385]: DEBUG oslo_vmware.api [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Task: {'id': task-4454531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21848} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.002022] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.002229] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1145.002400] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1145.002581] env[65385]: INFO nova.compute.manager [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1145.002834] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1145.003074] env[65385]: DEBUG nova.compute.manager [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1145.003190] env[65385]: DEBUG nova.network.neutron [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1145.003433] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.004076] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.004413] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.049707] env[65385]: DEBUG nova.network.neutron [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1145.050010] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.109218] env[65385]: DEBUG oslo_concurrency.lockutils [None req-11a726e4-a372-4fa5-a5a6-7cd9b85de96b tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.327s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.172776] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 3657ec21-3ec7-44c9-92d1-570655b58f26] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1145.292026] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.292158] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.293102] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a4b8da-d23e-419c-b08b-a0cd6415e847 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.314979] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fb14fb-7533-434d-9e4f-4c62dfcfc774 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.342020] env[65385]: WARNING neutronclient.v2_0.client [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.346716] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Reconfiguring VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1145.347091] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e2b4b32-b227-42eb-9f4f-a75bd8de2c24 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.370460] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1145.370460] env[65385]: value = "task-4454532" [ 1145.370460] env[65385]: _type = "Task" [ 1145.370460] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.383451] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.423028] env[65385]: DEBUG nova.compute.utils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1145.427074] env[65385]: DEBUG nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1145.427629] env[65385]: DEBUG nova.network.neutron [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1145.428238] env[65385]: WARNING neutronclient.v2_0.client [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.428724] env[65385]: WARNING neutronclient.v2_0.client [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.429568] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.430380] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.544293] env[65385]: DEBUG nova.policy [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662871f375e441eb98d814cf7d7ac5c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bc8afc699e34f059f18d8244980bc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1145.552972] env[65385]: DEBUG nova.network.neutron [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1145.677026] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 0673a0d2-76ac-4cd5-8cc3-b8596877c641] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1145.830367] env[65385]: DEBUG nova.compute.manager [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1145.833033] env[65385]: DEBUG nova.compute.manager [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing instance network info cache due to event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1145.833033] env[65385]: DEBUG oslo_concurrency.lockutils [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Acquiring lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.833033] env[65385]: DEBUG oslo_concurrency.lockutils [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Acquired lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.833033] env[65385]: DEBUG nova.network.neutron [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1145.882199] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.944709] env[65385]: DEBUG oslo_concurrency.lockutils [None req-20755817-1d15-44c3-a0df-442ee6d7f45e tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.027s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.948531] env[65385]: DEBUG nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1145.964578] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.974682] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c72e8a-5b8e-4d76-8a68-8adc5ac72bf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.999944] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance '5873d9d9-ff2d-4228-a4ed-b07717ba082e' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1146.009193] env[65385]: DEBUG nova.network.neutron [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Successfully created port: 4f519233-3caf-43f8-9330-571b1b60d4a3 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1146.041894] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.042766] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.044370] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.061423] env[65385]: INFO nova.compute.manager [-] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Took 1.06 seconds to deallocate network for instance. [ 1146.089767] env[65385]: DEBUG nova.network.neutron [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Port 92c45f41-92b4-40c3-8516-30189d29e33d binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1146.090067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.090278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.091457] env[65385]: DEBUG nova.network.neutron [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1146.183840] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 59f26752-75ab-45e5-b0a9-bbcef44ecd7f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1146.335794] env[65385]: WARNING neutronclient.v2_0.client [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.336504] env[65385]: WARNING openstack [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.336850] env[65385]: WARNING openstack [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.383691] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.506463] env[65385]: WARNING openstack [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.507010] env[65385]: WARNING openstack [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.516427] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.517820] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49ec4441-2fd4-478d-bc9d-d8d898a0d9ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.526645] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1146.526645] env[65385]: value = "task-4454533" [ 1146.526645] env[65385]: _type = "Task" [ 1146.526645] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.540996] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.567774] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1146.568106] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1146.568347] env[65385]: DEBUG nova.objects.instance [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lazy-loading 'resources' on Instance uuid 54faee8c-6b2d-4864-bf96-26b9dfba583b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.595232] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.596788] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.598046] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.659735] env[65385]: WARNING neutronclient.v2_0.client [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.660778] env[65385]: WARNING openstack [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.661020] env[65385]: WARNING openstack [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.686951] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 89f3c815-8671-47ce-9e74-bf6e652bb3c3] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1146.830389] env[65385]: DEBUG nova.network.neutron [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updated VIF entry in instance network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1146.830692] env[65385]: DEBUG nova.network.neutron [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf3df048-71", "ovs_interfaceid": "cf3df048-713a-474d-9278-a467e50ec4f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1146.866835] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.867315] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.723974] env[65385]: DEBUG nova.network.neutron [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Successfully updated port: 4f519233-3caf-43f8-9330-571b1b60d4a3 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1147.727615] env[65385]: DEBUG nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1147.732131] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 75ba706f-2e72-4f84-b02f-db4381951e77] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1147.734346] env[65385]: DEBUG oslo_concurrency.lockutils [req-bc839c29-9063-428e-b9d0-27f39dd1c953 req-7200718c-f947-401d-a7a4-c9131b859800 service nova] Releasing lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.735901] env[65385]: DEBUG nova.compute.manager [req-ade40274-2b5a-4f8f-a89b-78e4bf788849 req-90cb6931-f874-4a99-b89d-29c67f4e2f5b service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Received event network-vif-plugged-4f519233-3caf-43f8-9330-571b1b60d4a3 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1147.736217] env[65385]: DEBUG oslo_concurrency.lockutils [req-ade40274-2b5a-4f8f-a89b-78e4bf788849 req-90cb6931-f874-4a99-b89d-29c67f4e2f5b service nova] Acquiring lock "8db87973-e093-4da1-8534-62ea6896a57d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1147.736494] env[65385]: DEBUG oslo_concurrency.lockutils [req-ade40274-2b5a-4f8f-a89b-78e4bf788849 req-90cb6931-f874-4a99-b89d-29c67f4e2f5b service nova] Lock "8db87973-e093-4da1-8534-62ea6896a57d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.736736] env[65385]: DEBUG oslo_concurrency.lockutils [req-ade40274-2b5a-4f8f-a89b-78e4bf788849 req-90cb6931-f874-4a99-b89d-29c67f4e2f5b service nova] Lock "8db87973-e093-4da1-8534-62ea6896a57d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.736958] env[65385]: DEBUG nova.compute.manager [req-ade40274-2b5a-4f8f-a89b-78e4bf788849 req-90cb6931-f874-4a99-b89d-29c67f4e2f5b service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] No waiting events found dispatching network-vif-plugged-4f519233-3caf-43f8-9330-571b1b60d4a3 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1147.737154] env[65385]: WARNING nova.compute.manager [req-ade40274-2b5a-4f8f-a89b-78e4bf788849 req-90cb6931-f874-4a99-b89d-29c67f4e2f5b service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Received unexpected event network-vif-plugged-4f519233-3caf-43f8-9330-571b1b60d4a3 for instance with vm_state building and task_state spawning. [ 1147.738595] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "refresh_cache-8db87973-e093-4da1-8534-62ea6896a57d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.738720] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "refresh_cache-8db87973-e093-4da1-8534-62ea6896a57d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1147.739656] env[65385]: DEBUG nova.network.neutron [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1147.751740] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454533, 'name': PowerOffVM_Task, 'duration_secs': 0.336876} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.756841] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1147.757235] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance '5873d9d9-ff2d-4228-a4ed-b07717ba082e' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1147.760969] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.761245] env[65385]: WARNING oslo_vmware.common.loopingcall [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] task run outlasted interval by 0.385088 sec [ 1147.772646] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.779051] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.779051] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.779051] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.792200] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1147.792438] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1147.792659] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1147.792842] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1147.792994] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1147.793154] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1147.793355] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.793507] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1147.793689] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1147.793848] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1147.794026] env[65385]: DEBUG nova.virt.hardware [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1147.794899] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b4bf88-220a-4e8b-a41f-73799c1f29d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.804945] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5e20bc-dbdc-457b-93d6-1fe221e52f6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.889382] env[65385]: DEBUG nova.network.neutron [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1147.983436] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bddbb1-b125-4330-b040-2d18a5c8bfed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.991905] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc677b46-70b4-4651-82b8-c60067681c44 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.025018] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706f649a-33fa-4757-b39c-5c0e2c635c40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.033803] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd958f7-cd79-4e86-ac1e-3642a581c908 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.048580] env[65385]: DEBUG nova.compute.provider_tree [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.239494] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: eb9d0510-b453-4695-9e1d-731217b9f8ec] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1148.246719] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.247333] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.265175] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1148.265421] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1148.265563] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1148.265737] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1148.265885] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1148.266030] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1148.266234] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1148.266384] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1148.266539] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1148.266696] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1148.267055] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1148.272943] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6432c0fa-5bd5-45cb-8ff4-46035d109a36 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.291274] env[65385]: DEBUG nova.network.neutron [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1148.297458] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.300064] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1148.300064] env[65385]: value = "task-4454534" [ 1148.300064] env[65385]: _type = "Task" [ 1148.300064] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.310261] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454534, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.323469] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.324016] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.390233] env[65385]: WARNING neutronclient.v2_0.client [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.390958] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.391319] env[65385]: WARNING openstack [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.400181] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1148.481358] env[65385]: DEBUG nova.network.neutron [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Updating instance_info_cache with network_info: [{"id": "4f519233-3caf-43f8-9330-571b1b60d4a3", "address": "fa:16:3e:8a:71:7f", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f519233-3c", "ovs_interfaceid": "4f519233-3caf-43f8-9330-571b1b60d4a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1148.553115] env[65385]: DEBUG nova.scheduler.client.report [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.743278] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 59450963-d83f-46e8-8c13-05d4f1818c64] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1148.782523] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.810326] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454534, 'name': ReconfigVM_Task, 'duration_secs': 0.210044} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.810607] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance '5873d9d9-ff2d-4228-a4ed-b07717ba082e' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1148.904350] env[65385]: DEBUG nova.compute.manager [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65385) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:925}} [ 1148.904598] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.984305] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "refresh_cache-8db87973-e093-4da1-8534-62ea6896a57d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1148.985141] env[65385]: DEBUG nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Instance network_info: |[{"id": "4f519233-3caf-43f8-9330-571b1b60d4a3", "address": "fa:16:3e:8a:71:7f", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f519233-3c", "ovs_interfaceid": "4f519233-3caf-43f8-9330-571b1b60d4a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1148.985778] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:71:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f519233-3caf-43f8-9330-571b1b60d4a3', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1148.993638] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1148.993930] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1148.994186] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f42578ea-f7e3-4404-ba22-ce50f2414997 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.018939] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1149.018939] env[65385]: value = "task-4454535" [ 1149.018939] env[65385]: _type = "Task" [ 1149.018939] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.029029] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454535, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.059223] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.491s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.061721] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.157s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.084142] env[65385]: INFO nova.scheduler.client.report [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Deleted allocations for instance 54faee8c-6b2d-4864-bf96-26b9dfba583b [ 1149.247045] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 0115dde8-b0aa-4960-a452-ced6587bc567] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1149.283423] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.317409] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1149.317609] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1149.317734] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1149.317953] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1149.318124] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1149.318266] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1149.318466] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1149.318614] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1149.318771] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1149.318928] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1149.319106] env[65385]: DEBUG nova.virt.hardware [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1149.326704] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1149.327070] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c39dded-4842-4211-a838-73ba2b0dcf2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.347915] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1149.347915] env[65385]: value = "task-4454536" [ 1149.347915] env[65385]: _type = "Task" [ 1149.347915] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.357093] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454536, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.530478] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454535, 'name': CreateVM_Task, 'duration_secs': 0.479108} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.530676] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1149.531300] env[65385]: WARNING neutronclient.v2_0.client [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1149.532026] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.532026] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1149.532254] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1149.532448] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11d4cb53-efec-4767-a9c0-a8d17cffbd2d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.539185] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1149.539185] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9e8bb-ebf0-bcd3-97be-afe2bc046d43" [ 1149.539185] env[65385]: _type = "Task" [ 1149.539185] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.549788] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9e8bb-ebf0-bcd3-97be-afe2bc046d43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.559062] env[65385]: DEBUG nova.compute.manager [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Received event network-changed-4f519233-3caf-43f8-9330-571b1b60d4a3 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1149.559287] env[65385]: DEBUG nova.compute.manager [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Refreshing instance network info cache due to event network-changed-4f519233-3caf-43f8-9330-571b1b60d4a3. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1149.559497] env[65385]: DEBUG oslo_concurrency.lockutils [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Acquiring lock "refresh_cache-8db87973-e093-4da1-8534-62ea6896a57d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.559630] env[65385]: DEBUG oslo_concurrency.lockutils [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Acquired lock "refresh_cache-8db87973-e093-4da1-8534-62ea6896a57d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1149.559785] env[65385]: DEBUG nova.network.neutron [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Refreshing network info cache for port 4f519233-3caf-43f8-9330-571b1b60d4a3 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1149.564163] env[65385]: DEBUG nova.objects.instance [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'migration_context' on Instance uuid 6c826914-4c6b-456b-a403-2373d30c803f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.594013] env[65385]: DEBUG oslo_concurrency.lockutils [None req-297d7c3d-850b-4982-912f-9861e78097a4 tempest-ServerShowV254Test-503814553 tempest-ServerShowV254Test-503814553-project-member] Lock "54faee8c-6b2d-4864-bf96-26b9dfba583b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.270s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.752832] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 42b9edf6-1873-49c3-8074-8eef654ac371] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1149.783562] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.859889] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454536, 'name': ReconfigVM_Task, 'duration_secs': 0.259511} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.860171] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1149.861068] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfbdbc0-1187-486f-8287-8c213ba33a30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.886493] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 5873d9d9-ff2d-4228-a4ed-b07717ba082e/5873d9d9-ff2d-4228-a4ed-b07717ba082e.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.887019] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b9a0188-501c-40c6-bf4e-60212959667b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.906261] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1149.906261] env[65385]: value = "task-4454537" [ 1149.906261] env[65385]: _type = "Task" [ 1149.906261] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.917226] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454537, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.052944] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9e8bb-ebf0-bcd3-97be-afe2bc046d43, 'name': SearchDatastore_Task, 'duration_secs': 0.024354} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.053374] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1150.054044] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1150.054044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.054044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.054246] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1150.054591] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a379681-739c-4479-99c6-f7cbf9c36998 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.062711] env[65385]: WARNING neutronclient.v2_0.client [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.063632] env[65385]: WARNING openstack [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.064081] env[65385]: WARNING openstack [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.075964] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1150.077153] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1150.077468] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3ec80c7-22ad-4926-918a-aa1953f65f61 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.085348] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1150.085348] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52acebbe-9c50-ae26-3117-ec862bdcbfa5" [ 1150.085348] env[65385]: _type = "Task" [ 1150.085348] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.096819] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52acebbe-9c50-ae26-3117-ec862bdcbfa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.172863] env[65385]: WARNING openstack [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.173362] env[65385]: WARNING openstack [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.256565] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.256846] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Cleaning up deleted instances with incomplete migration {{(pid=65385) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11954}} [ 1150.259331] env[65385]: WARNING neutronclient.v2_0.client [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.259711] env[65385]: WARNING openstack [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.260085] env[65385]: WARNING openstack [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.286148] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.301772] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7187d1a-2f0d-4736-987f-f5b19e1b1219 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.310628] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669ee89c-529f-4b83-b623-62ce5f349e36 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.347221] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38fca2e-3cf4-43fd-b4fb-0c6564dca800 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.356248] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808e6ce1-8793-4c68-9c43-ab571d4374fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.362728] env[65385]: DEBUG nova.network.neutron [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Updated VIF entry in instance network info cache for port 4f519233-3caf-43f8-9330-571b1b60d4a3. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1150.363106] env[65385]: DEBUG nova.network.neutron [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Updating instance_info_cache with network_info: [{"id": "4f519233-3caf-43f8-9330-571b1b60d4a3", "address": "fa:16:3e:8a:71:7f", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f519233-3c", "ovs_interfaceid": "4f519233-3caf-43f8-9330-571b1b60d4a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1150.375387] env[65385]: DEBUG nova.compute.provider_tree [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.418979] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454537, 'name': ReconfigVM_Task, 'duration_secs': 0.40796} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.419276] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 5873d9d9-ff2d-4228-a4ed-b07717ba082e/5873d9d9-ff2d-4228-a4ed-b07717ba082e.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.419525] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance '5873d9d9-ff2d-4228-a4ed-b07717ba082e' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1150.597988] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52acebbe-9c50-ae26-3117-ec862bdcbfa5, 'name': SearchDatastore_Task, 'duration_secs': 0.014892} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.598934] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-977ac18e-8dd5-4300-be7f-013a50395f1b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.606103] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1150.606103] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526ca552-17d1-01a3-c1c2-22050d5cc7aa" [ 1150.606103] env[65385]: _type = "Task" [ 1150.606103] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.615594] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526ca552-17d1-01a3-c1c2-22050d5cc7aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.784537] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.866712] env[65385]: DEBUG oslo_concurrency.lockutils [req-dfa1d424-c66f-4770-9561-8aac537ac165 req-e60d3046-ab3a-422f-80da-b8b02ed06914 service nova] Releasing lock "refresh_cache-8db87973-e093-4da1-8534-62ea6896a57d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1150.879250] env[65385]: DEBUG nova.scheduler.client.report [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1150.925988] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d75fb5-5930-4e9f-b528-d64ac0da3a3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.946612] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96ccb42-9234-4a96-bedf-f28ff8de6c55 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.966476] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance '5873d9d9-ff2d-4228-a4ed-b07717ba082e' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1151.117497] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]526ca552-17d1-01a3-c1c2-22050d5cc7aa, 'name': SearchDatastore_Task, 'duration_secs': 0.015338} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.119072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.119072] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 8db87973-e093-4da1-8534-62ea6896a57d/8db87973-e093-4da1-8534-62ea6896a57d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1151.119072] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14ed0368-2f1d-4f84-bc46-766905a34eb4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.127194] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1151.127194] env[65385]: value = "task-4454538" [ 1151.127194] env[65385]: _type = "Task" [ 1151.127194] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.136174] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.285232] env[65385]: DEBUG oslo_vmware.api [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454532, 'name': ReconfigVM_Task, 'duration_secs': 5.772379} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.285609] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.285900] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Reconfigured VM to detach interface {{(pid=65385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1151.286414] env[65385]: WARNING neutronclient.v2_0.client [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.286788] env[65385]: WARNING neutronclient.v2_0.client [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.287482] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.287808] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.326242] env[65385]: WARNING neutronclient.v2_0.client [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.474075] env[65385]: WARNING neutronclient.v2_0.client [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.474270] env[65385]: WARNING neutronclient.v2_0.client [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.515627] env[65385]: DEBUG nova.network.neutron [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Port 7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1151.641990] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454538, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.892571] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.831s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.141132] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.763254} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.141499] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 8db87973-e093-4da1-8534-62ea6896a57d/8db87973-e093-4da1-8534-62ea6896a57d.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1152.141701] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1152.141982] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2b0d3ea-3e6d-46b6-aa78-fed3d87ac5de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.151512] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1152.151512] env[65385]: value = "task-4454539" [ 1152.151512] env[65385]: _type = "Task" [ 1152.151512] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.163365] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.329393] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.329841] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.330030] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.330211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.330385] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.333213] env[65385]: INFO nova.compute.manager [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Terminating instance [ 1152.541043] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.541251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.541404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.646475] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.646716] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquired lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.646984] env[65385]: DEBUG nova.network.neutron [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1152.663243] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112742} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.663522] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1152.664451] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad12c79-ea0d-4738-9cd1-e7b061fd5d29 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.692308] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 8db87973-e093-4da1-8534-62ea6896a57d/8db87973-e093-4da1-8534-62ea6896a57d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.692727] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab2e0770-8f71-443d-8cdb-a3b99bf3a0d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.716619] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1152.716619] env[65385]: value = "task-4454540" [ 1152.716619] env[65385]: _type = "Task" [ 1152.716619] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.727332] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454540, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.769791] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1152.769791] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1152.838016] env[65385]: DEBUG nova.compute.manager [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1152.838279] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1152.839172] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800d6826-5ee6-4ed0-9359-0f6ebb3a60d0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.847894] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1152.848231] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-011c782b-a246-4192-a27c-4e331f8b0a90 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.856451] env[65385]: DEBUG oslo_vmware.api [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1152.856451] env[65385]: value = "task-4454541" [ 1152.856451] env[65385]: _type = "Task" [ 1152.856451] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.868498] env[65385]: DEBUG oslo_vmware.api [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454541, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.921702] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.922124] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.150624] env[65385]: WARNING neutronclient.v2_0.client [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1153.151480] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.151850] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.228020] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454540, 'name': ReconfigVM_Task, 'duration_secs': 0.381244} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.228020] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 8db87973-e093-4da1-8534-62ea6896a57d/8db87973-e093-4da1-8534-62ea6896a57d.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1153.228287] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f391d4c-2550-4815-9d8c-73aa98fad19f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.238923] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1153.238923] env[65385]: value = "task-4454542" [ 1153.238923] env[65385]: _type = "Task" [ 1153.238923] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.251864] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454542, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.277425] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.280295] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.280755] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.292127] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.292702] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.292936] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.293168] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_power_states {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.355360] env[65385]: WARNING neutronclient.v2_0.client [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1153.356128] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.356488] env[65385]: WARNING openstack [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.377064] env[65385]: DEBUG oslo_vmware.api [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454541, 'name': PowerOffVM_Task, 'duration_secs': 0.29504} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.377064] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1153.377064] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1153.377336] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b81d9f68-cb31-4e05-a417-525415175a20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.425744] env[65385]: DEBUG nova.compute.utils [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1153.438120] env[65385]: INFO nova.compute.manager [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Swapping old allocation on dict_keys(['1af23b69-5ce6-4d6c-8591-1b95ecca8a6b']) held by migration 77bd2d4a-6632-4319-9c82-871dbac8ac69 for instance [ 1153.448613] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1153.448985] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1153.449296] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleting the datastore file [datastore1] b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.449692] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65f0a2de-eeca-4f45-994d-75aff0c8e041 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.456727] env[65385]: INFO nova.network.neutron [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Port 330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1153.457096] env[65385]: DEBUG nova.network.neutron [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [{"id": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "address": "fa:16:3e:09:36:60", "network": {"id": "16967e55-c149-4a58-905f-c47b7facac11", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-320100029-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1991528775bd4557ae6150b30095e923", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e6c199b-75", "ovs_interfaceid": "6e6c199b-75f0-473b-90b5-0daaf633ba14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1153.464705] env[65385]: DEBUG oslo_vmware.api [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1153.464705] env[65385]: value = "task-4454544" [ 1153.464705] env[65385]: _type = "Task" [ 1153.464705] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.469195] env[65385]: DEBUG nova.scheduler.client.report [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Overwriting current allocation {'allocations': {'1af23b69-5ce6-4d6c-8591-1b95ecca8a6b': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 158}}, 'project_id': '60de0d1c162342209795a180391661f6', 'user_id': 'b2a287457352470794887b229ea90cab', 'consumer_generation': 1} on consumer 6c826914-4c6b-456b-a403-2373d30c803f {{(pid=65385) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1153.479635] env[65385]: DEBUG oslo_vmware.api [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454544, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.532652] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1153.548063] env[65385]: WARNING neutronclient.v2_0.client [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1153.567893] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.567893] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.567893] env[65385]: DEBUG nova.network.neutron [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1153.581816] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.581970] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.582164] env[65385]: DEBUG nova.network.neutron [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1153.749607] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454542, 'name': Rename_Task, 'duration_secs': 0.165157} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.749818] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1153.750086] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edfc03b8-9ed3-4990-b534-f13bb37d19a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.758069] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1153.758069] env[65385]: value = "task-4454545" [ 1153.758069] env[65385]: _type = "Task" [ 1153.758069] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.767048] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454545, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.799658] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Getting list of instances from cluster (obj){ [ 1153.799658] env[65385]: value = "domain-c8" [ 1153.799658] env[65385]: _type = "ClusterComputeResource" [ 1153.799658] env[65385]: } {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1153.801211] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eb2ad5-29d4-4781-89d1-57ec5919efcc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.822872] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Got total of 10 instances {{(pid=65385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1153.823067] env[65385]: WARNING nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] While synchronizing instance power states, found 11 instances in the database and 10 instances on the hypervisor. [ 1153.823228] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid b4bca2c8-fc8f-471a-899b-aac648fee9fe {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.823669] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid 992920c3-af4f-489e-b21f-9d52c24399de {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.823912] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid 900639b6-9b98-436f-aaad-ea8391f67393 {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.824148] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid 6c826914-4c6b-456b-a403-2373d30c803f {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.824386] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.824728] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.824932] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid b92bc73a-9666-435e-b387-18df7c838f9b {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.825195] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid 62c30ded-c851-477f-b1d4-921268a6ef1c {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.825395] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid 5873d9d9-ff2d-4228-a4ed-b07717ba082e {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.825589] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.825784] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Triggering sync for uuid 8db87973-e093-4da1-8534-62ea6896a57d {{(pid=65385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1153.826366] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.826611] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.826899] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "992920c3-af4f-489e-b21f-9d52c24399de" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.827092] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "992920c3-af4f-489e-b21f-9d52c24399de" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.827335] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.827504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "900639b6-9b98-436f-aaad-ea8391f67393" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.827831] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "6c826914-4c6b-456b-a403-2373d30c803f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.828081] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "6c826914-4c6b-456b-a403-2373d30c803f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.828256] env[65385]: INFO nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1153.828419] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "6c826914-4c6b-456b-a403-2373d30c803f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.828695] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.828911] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.829131] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.829396] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.829585] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.829836] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.830023] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.830188] env[65385]: INFO nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] During sync_power_state the instance has a pending task (resize_migrated). Skip. [ 1153.830343] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.830582] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.830810] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "20933973-82ae-498c-b016-3a82bb5dc165" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.831074] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "8db87973-e093-4da1-8534-62ea6896a57d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.831267] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.831400] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1153.832235] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781aa64f-bcc0-4fe1-baf4-3c6dae30e88a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.835799] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7f10d8-5f12-4bbc-97e9-e4820c13fcf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.838787] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602fb69f-193a-4c6e-82af-daa2315be621 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.841826] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd29913-d129-406b-8760-a77fe823f4ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.844671] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6731513d-f308-4840-9050-f60c5cfcdcf6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.847188] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.929769] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.930197] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.101s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.931146] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdec62c-4671-480e-b663-3d56033c6681 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.960168] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Releasing lock "refresh_cache-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.975714] env[65385]: DEBUG oslo_vmware.api [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454544, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.434159} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.976021] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.976222] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1153.976417] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1153.976587] env[65385]: INFO nova.compute.manager [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1153.976839] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1153.977112] env[65385]: DEBUG nova.compute.manager [-] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1153.977233] env[65385]: DEBUG nova.network.neutron [-] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1153.977493] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1153.978084] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.978371] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.018413] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.072032] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.072721] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.073160] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.085217] env[65385]: WARNING neutronclient.v2_0.client [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.085952] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.086276] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.269719] env[65385]: DEBUG oslo_vmware.api [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454545, 'name': PowerOnVM_Task, 'duration_secs': 0.477801} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.270084] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1154.270353] env[65385]: INFO nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Took 6.54 seconds to spawn the instance on the hypervisor. [ 1154.270559] env[65385]: DEBUG nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1154.271443] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e2b57a-6e7b-4bf1-9ed7-9d14dfab28c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.284907] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.285458] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.300813] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.301441] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.353387] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.353707] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.353915] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.354151] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1154.355448] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3b0ce3-6af7-4a0c-b994-0838e8a3f0a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.360676] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "900639b6-9b98-436f-aaad-ea8391f67393" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.533s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.368079] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.541s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.368593] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "992920c3-af4f-489e-b21f-9d52c24399de" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.541s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.368943] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "20933973-82ae-498c-b016-3a82bb5dc165" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.538s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.370417] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db85293-9d7a-4bbf-9386-462e582f6ade {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.374759] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.545s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.386523] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39390654-292d-4ee5-9c16-9a73d45aa4d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.394792] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ff93fa-470f-452f-8d6d-dc5e493255f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.399202] env[65385]: WARNING neutronclient.v2_0.client [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.399850] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.400202] env[65385]: WARNING openstack [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.438901] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178976MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1154.439097] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.439942] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.443058] env[65385]: WARNING neutronclient.v2_0.client [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.443638] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.444235] env[65385]: WARNING openstack [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.452616] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.466063] env[65385]: DEBUG oslo_concurrency.lockutils [None req-eda55933-ab7c-476e-be97-77543c368475 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "interface-b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f-330c3a6f-16e2-4e4d-ae21-11b0c5b0dce1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.677s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.537253] env[65385]: DEBUG nova.network.neutron [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [{"id": "92c45f41-92b4-40c3-8516-30189d29e33d", "address": "fa:16:3e:78:19:01", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c45f41-92", "ovs_interfaceid": "92c45f41-92b4-40c3-8516-30189d29e33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.571300] env[65385]: DEBUG nova.network.neutron [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance_info_cache with network_info: [{"id": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "address": "fa:16:3e:b0:35:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e4dcc22-4c", "ovs_interfaceid": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.791220] env[65385]: INFO nova.compute.manager [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Took 13.87 seconds to build instance. [ 1155.008843] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1155.009227] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.009502] env[65385]: INFO nova.compute.manager [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Attaching volume fd0b7de1-8094-45a0-a553-bc8d5424ff10 to /dev/sdb [ 1155.037112] env[65385]: DEBUG oslo_concurrency.lockutils [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-6c826914-4c6b-456b-a403-2373d30c803f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.037601] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1155.037924] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3714ceff-731a-4ba4-b17c-02c525554c49 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.050632] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1155.050632] env[65385]: value = "task-4454546" [ 1155.050632] env[65385]: _type = "Task" [ 1155.050632] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.062270] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc100af3-b923-4e62-bb98-80b164170105 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.070952] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454546, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.073375] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77626b73-f661-4fb6-8cf9-4673424bd362 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.076746] env[65385]: DEBUG oslo_concurrency.lockutils [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.095784] env[65385]: DEBUG nova.virt.block_device [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updating existing volume attachment record: e9659b0f-e060-49b9-956c-325ec8d9c2aa {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1155.196822] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3a040-c133-18d4-babf-590d1ec542ca/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1155.197422] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951c8e9a-45a5-4edf-bc6a-785820d2669f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.205380] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3a040-c133-18d4-babf-590d1ec542ca/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1155.205564] env[65385]: ERROR oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3a040-c133-18d4-babf-590d1ec542ca/disk-0.vmdk due to incomplete transfer. [ 1155.205876] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7b30cadb-20f5-4ba9-a8b3-5e78ab069e21 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.214966] env[65385]: DEBUG oslo_vmware.rw_handles [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3a040-c133-18d4-babf-590d1ec542ca/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1155.215197] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Uploaded image 1b100802-b12e-4f1f-a998-6080dbb6821d to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1155.218405] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1155.218724] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2a70b9aa-3ec4-47c4-b6ad-885c59dc0c7e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.225414] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1155.225414] env[65385]: value = "task-4454547" [ 1155.225414] env[65385]: _type = "Task" [ 1155.225414] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.239198] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454547, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.253022] env[65385]: DEBUG nova.compute.manager [req-9a6976c8-5802-42ed-ab80-482edf119934 req-c151cc29-01cd-4e72-80de-fef7d970be0b service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Received event network-vif-deleted-6e6c199b-75f0-473b-90b5-0daaf633ba14 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1155.253146] env[65385]: INFO nova.compute.manager [req-9a6976c8-5802-42ed-ab80-482edf119934 req-c151cc29-01cd-4e72-80de-fef7d970be0b service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Neutron deleted interface 6e6c199b-75f0-473b-90b5-0daaf633ba14; detaching it from the instance and deleting it from the info cache [ 1155.253302] env[65385]: DEBUG nova.network.neutron [req-9a6976c8-5802-42ed-ab80-482edf119934 req-c151cc29-01cd-4e72-80de-fef7d970be0b service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1155.293789] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2c9a8eb7-c0ae-4e64-b904-6ea0999e1508 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "8db87973-e093-4da1-8534-62ea6896a57d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.380s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.294230] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "8db87973-e093-4da1-8534-62ea6896a57d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.463s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.294373] env[65385]: INFO nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] During sync_power_state the instance has a pending task (spawning). Skip. [ 1155.294540] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "8db87973-e093-4da1-8534-62ea6896a57d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.467317] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Applying migration context for instance 5873d9d9-ff2d-4228-a4ed-b07717ba082e as it has an incoming, in-progress migration 63618a0e-63c2-49ca-84d5-466f3f3f0c0d. Migration status is post-migrating {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1155.468814] env[65385]: INFO nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating resource usage from migration 63618a0e-63c2-49ca-84d5-466f3f3f0c0d [ 1155.491799] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b4bca2c8-fc8f-471a-899b-aac648fee9fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.491971] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 992920c3-af4f-489e-b21f-9d52c24399de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492106] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 900639b6-9b98-436f-aaad-ea8391f67393 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492262] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492347] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance df419705-e7e8-47b4-b9a0-2f1cf2638f33 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492468] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b92bc73a-9666-435e-b387-18df7c838f9b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492574] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 62c30ded-c851-477f-b1d4-921268a6ef1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492674] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 6c826914-4c6b-456b-a403-2373d30c803f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492776] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 20933973-82ae-498c-b016-3a82bb5dc165 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.492879] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Migration 63618a0e-63c2-49ca-84d5-466f3f3f0c0d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1155.492980] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 5873d9d9-ff2d-4228-a4ed-b07717ba082e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.493094] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 8db87973-e093-4da1-8534-62ea6896a57d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1155.493296] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1155.493448] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=100GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '11', 'num_vm_active': '10', 'num_task_None': '7', 'num_os_type_None': '11', 'num_proj_8d45ff3446a84ef2a1945f7a66b747e8': '2', 'io_workload': '0', 'num_proj_6bc8afc699e34f059f18d8244980bc1d': '2', 'num_proj_1991528775bd4557ae6150b30095e923': '2', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_60de0d1c162342209795a180391661f6': '1', 'num_task_deleting': '1', 'num_task_shelving_image_uploading': '1', 'num_proj_b27c092a0fd345ea953956469c21f8bb': '1', 'num_proj_2ba61d77cc2d4fe1ba3c03466d5985d9': '1', 'num_task_resize_migrated': '1', 'num_proj_94cc5bc221c0455ea760f5022db6bdc4': '1', 'num_proj_1598307c304d46a884a88b3eb5f5d7dc': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1155.565407] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454546, 'name': PowerOffVM_Task, 'duration_secs': 0.337243} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.568647] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1155.570046] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1155.570046] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1155.570192] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1155.570375] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1155.570527] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1155.570671] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1155.570897] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1155.571097] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1155.571327] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1155.571516] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1155.571689] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1155.577399] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe5febc3-54e9-4975-9bfd-c9223618b9b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.605947] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1155.605947] env[65385]: value = "task-4454549" [ 1155.605947] env[65385]: _type = "Task" [ 1155.605947] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.621886] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa73be9-a77f-460d-bad6-67b99a53c11c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.625961] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454549, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.647172] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5381a5e-b416-4e74-aef6-7981623259b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.658184] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance '5873d9d9-ff2d-4228-a4ed-b07717ba082e' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1155.728355] env[65385]: DEBUG nova.network.neutron [-] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1155.730355] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd385ee-8f97-4c14-9aee-e87bb67334ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.745030] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ae70b4-8973-48f6-b991-41094b1e34fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.749144] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454547, 'name': Destroy_Task, 'duration_secs': 0.399978} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.749144] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Destroyed the VM [ 1155.749144] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1155.749668] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cd437a7e-60dd-4e2b-8d42-b831e33de6f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.786860] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7db37b17-98e2-486a-94b7-9383af4d6a8f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.789740] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2326aff0-4fcc-4f48-bcb7-b0b9b5c2c28e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.791849] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1155.791849] env[65385]: value = "task-4454550" [ 1155.791849] env[65385]: _type = "Task" [ 1155.791849] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.801849] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad99e51-8173-424c-ac35-65e301195533 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.812372] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7045429f-30c4-4187-9c15-375452273097 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.823160] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454550, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.834967] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.850218] env[65385]: DEBUG nova.compute.manager [req-9a6976c8-5802-42ed-ab80-482edf119934 req-c151cc29-01cd-4e72-80de-fef7d970be0b service nova] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Detach interface failed, port_id=6e6c199b-75f0-473b-90b5-0daaf633ba14, reason: Instance b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1156.117593] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454549, 'name': ReconfigVM_Task, 'duration_secs': 0.273398} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.118436] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e506eb76-fb83-4eee-b6a2-4d1a012b22f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.140586] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1156.140905] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1156.141122] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1156.141348] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1156.141545] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1156.141725] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1156.141987] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.142212] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1156.142442] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1156.142646] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1156.142854] env[65385]: DEBUG nova.virt.hardware [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1156.143752] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-480f84b0-2c9e-4b24-9d72-cc4121cfeb22 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.151516] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1156.151516] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5205e82d-50c7-d386-6c2e-76e462ae2b10" [ 1156.151516] env[65385]: _type = "Task" [ 1156.151516] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.155334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "8db87973-e093-4da1-8534-62ea6896a57d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.155613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "8db87973-e093-4da1-8534-62ea6896a57d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.155823] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "8db87973-e093-4da1-8534-62ea6896a57d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.156643] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "8db87973-e093-4da1-8534-62ea6896a57d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.156643] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "8db87973-e093-4da1-8534-62ea6896a57d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.158458] env[65385]: INFO nova.compute.manager [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Terminating instance [ 1156.166525] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1156.172231] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0e29d99-c29b-4f7e-8fc1-0a93d7f02db6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.174885] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5205e82d-50c7-d386-6c2e-76e462ae2b10, 'name': SearchDatastore_Task, 'duration_secs': 0.014609} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.182508] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1156.184011] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff720db3-7f5b-4da8-a974-028a4b0147fd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.202093] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1156.202093] env[65385]: value = "task-4454551" [ 1156.202093] env[65385]: _type = "Task" [ 1156.202093] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.207800] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1156.207800] env[65385]: value = "task-4454552" [ 1156.207800] env[65385]: _type = "Task" [ 1156.207800] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.211364] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454551, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.222066] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454552, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.234434] env[65385]: INFO nova.compute.manager [-] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Took 2.26 seconds to deallocate network for instance. [ 1156.302384] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454550, 'name': RemoveSnapshot_Task, 'duration_secs': 0.426636} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.302741] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1156.303162] env[65385]: DEBUG nova.compute.manager [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1156.303974] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd869c81-1d07-4f15-8d8a-277ccd06fc3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.344597] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1156.662499] env[65385]: DEBUG nova.compute.manager [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1156.662737] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1156.663755] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e761ba-db44-405a-a6cb-a6559e9d1a87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.672685] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1156.672954] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4a0096c-a701-4a32-b9e5-1dac04ead8a5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.680498] env[65385]: DEBUG oslo_vmware.api [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1156.680498] env[65385]: value = "task-4454553" [ 1156.680498] env[65385]: _type = "Task" [ 1156.680498] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.690204] env[65385]: DEBUG oslo_vmware.api [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.713861] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454551, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.723929] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454552, 'name': ReconfigVM_Task, 'duration_secs': 0.306759} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.724214] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1156.725102] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84319fe-43b2-4cd6-a869-91c32d23f37a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.742904] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.751213] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.751643] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d656e93-0eaa-41ce-91c1-03dd58a657e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.772348] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1156.772348] env[65385]: value = "task-4454554" [ 1156.772348] env[65385]: _type = "Task" [ 1156.772348] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.782329] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454554, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.818160] env[65385]: INFO nova.compute.manager [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Shelve offloading [ 1156.853091] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1156.853301] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.414s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.853607] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.111s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.853861] env[65385]: DEBUG nova.objects.instance [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'resources' on Instance uuid b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.192488] env[65385]: DEBUG oslo_vmware.api [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454553, 'name': PowerOffVM_Task, 'duration_secs': 0.236549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.192856] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1157.192895] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1157.193210] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f3c31c6-630c-45ea-83ea-68e4bf43b94f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.212167] env[65385]: DEBUG oslo_vmware.api [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454551, 'name': PowerOnVM_Task, 'duration_secs': 0.623308} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.212425] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1157.212596] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-12578204-cb98-4d3e-a979-b76e77d7a7de tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance '5873d9d9-ff2d-4228-a4ed-b07717ba082e' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1157.262875] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1157.263226] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1157.263575] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleting the datastore file [datastore2] 8db87973-e093-4da1-8534-62ea6896a57d {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1157.264020] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc20bf21-1d3b-418e-9a2b-919e432b2eb5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.274683] env[65385]: DEBUG oslo_vmware.api [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1157.274683] env[65385]: value = "task-4454556" [ 1157.274683] env[65385]: _type = "Task" [ 1157.274683] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.293328] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.298862] env[65385]: DEBUG oslo_vmware.api [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.324367] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.325112] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c46df7d-f85c-48e2-beee-c1223bd3d17a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.334031] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1157.334031] env[65385]: value = "task-4454557" [ 1157.334031] env[65385]: _type = "Task" [ 1157.334031] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.347854] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1157.348224] env[65385]: DEBUG nova.compute.manager [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1157.349484] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfa6af2-bdb3-49bf-ba4c-a5b1e94cdb9f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.363276] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.363612] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1157.363841] env[65385]: DEBUG nova.network.neutron [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1157.525636] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0847e1c-63e3-48fa-a4fb-b039a6ad897f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.534228] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0674ffc-c332-4ea3-9be3-d8254c4bdde6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.565131] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567c4427-0b9f-4c4e-bcd9-26acf5e3f3ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.573656] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227290c6-d9b7-4382-935a-b65ac7ca050a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.589369] env[65385]: DEBUG nova.compute.provider_tree [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.786235] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454554, 'name': ReconfigVM_Task, 'duration_secs': 0.645052} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.786838] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f/6c826914-4c6b-456b-a403-2373d30c803f.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1157.787701] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5fb25e-ddea-4503-93c7-477f842a8d19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.793031] env[65385]: DEBUG oslo_vmware.api [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454556, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206201} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.793575] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1157.793761] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1157.793935] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1157.794145] env[65385]: INFO nova.compute.manager [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1157.794357] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1157.794570] env[65385]: DEBUG nova.compute.manager [-] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1157.794692] env[65385]: DEBUG nova.network.neutron [-] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1157.794949] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.795497] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.795757] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.816765] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693a366f-6c62-44ec-8048-494110a32e3d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.835507] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9587ce69-6929-4ed1-a972-f9229e3530e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.853691] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0caa6240-cc97-4be1-813f-d50bc50bfc40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.857223] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.863784] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1157.864041] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbc5aace-284b-44bf-a765-bd34852635f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.866261] env[65385]: WARNING neutronclient.v2_0.client [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.866941] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.867479] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.875604] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1157.875604] env[65385]: value = "task-4454559" [ 1157.875604] env[65385]: _type = "Task" [ 1157.875604] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.884686] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454559, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.021710] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.022465] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.093729] env[65385]: DEBUG nova.scheduler.client.report [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1158.141711] env[65385]: DEBUG nova.compute.manager [req-7c30cfde-1a66-4607-abab-740c55efee48 req-cc2cc3d7-678a-45c5-bf24-5464d1d68d6d service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Received event network-vif-deleted-4f519233-3caf-43f8-9330-571b1b60d4a3 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1158.141903] env[65385]: INFO nova.compute.manager [req-7c30cfde-1a66-4607-abab-740c55efee48 req-cc2cc3d7-678a-45c5-bf24-5464d1d68d6d service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Neutron deleted interface 4f519233-3caf-43f8-9330-571b1b60d4a3; detaching it from the instance and deleting it from the info cache [ 1158.142085] env[65385]: DEBUG nova.network.neutron [req-7c30cfde-1a66-4607-abab-740c55efee48 req-cc2cc3d7-678a-45c5-bf24-5464d1d68d6d service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1158.162178] env[65385]: WARNING neutronclient.v2_0.client [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.162948] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.163543] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.269757] env[65385]: DEBUG nova.network.neutron [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd731def-a6", "ovs_interfaceid": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1158.388643] env[65385]: DEBUG oslo_vmware.api [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454559, 'name': PowerOnVM_Task, 'duration_secs': 0.417576} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.389011] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1158.599193] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.612198] env[65385]: DEBUG nova.network.neutron [-] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1158.625534] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.625900] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1158.628608] env[65385]: INFO nova.scheduler.client.report [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted allocations for instance b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f [ 1158.645361] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50a77f52-991f-4b45-bfee-146737cd4d79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.657154] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbd7eae-4259-450d-887b-344b0e1ff20f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.694978] env[65385]: DEBUG nova.compute.manager [req-7c30cfde-1a66-4607-abab-740c55efee48 req-cc2cc3d7-678a-45c5-bf24-5464d1d68d6d service nova] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Detach interface failed, port_id=4f519233-3caf-43f8-9330-571b1b60d4a3, reason: Instance 8db87973-e093-4da1-8534-62ea6896a57d could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1158.772328] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1158.772767] env[65385]: WARNING neutronclient.v2_0.client [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.773449] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.773842] env[65385]: WARNING openstack [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.779000] env[65385]: WARNING neutronclient.v2_0.client [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.069517] env[65385]: DEBUG nova.compute.manager [req-a1d2c94f-7697-4c71-be33-6be93594c5ec req-c88804e0-6dcb-4ae5-b890-236ea656a8da service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-vif-unplugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1159.069740] env[65385]: DEBUG oslo_concurrency.lockutils [req-a1d2c94f-7697-4c71-be33-6be93594c5ec req-c88804e0-6dcb-4ae5-b890-236ea656a8da service nova] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.070054] env[65385]: DEBUG oslo_concurrency.lockutils [req-a1d2c94f-7697-4c71-be33-6be93594c5ec req-c88804e0-6dcb-4ae5-b890-236ea656a8da service nova] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.070164] env[65385]: DEBUG oslo_concurrency.lockutils [req-a1d2c94f-7697-4c71-be33-6be93594c5ec req-c88804e0-6dcb-4ae5-b890-236ea656a8da service nova] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.070278] env[65385]: DEBUG nova.compute.manager [req-a1d2c94f-7697-4c71-be33-6be93594c5ec req-c88804e0-6dcb-4ae5-b890-236ea656a8da service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] No waiting events found dispatching network-vif-unplugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1159.070451] env[65385]: WARNING nova.compute.manager [req-a1d2c94f-7697-4c71-be33-6be93594c5ec req-c88804e0-6dcb-4ae5-b890-236ea656a8da service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received unexpected event network-vif-unplugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 for instance with vm_state shelved and task_state shelving_offloading. [ 1159.116763] env[65385]: INFO nova.compute.manager [-] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Took 1.32 seconds to deallocate network for instance. [ 1159.129540] env[65385]: DEBUG nova.compute.utils [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1159.139979] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9de108d8-0852-44d5-9fc0-a1b6be4a7d73 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.810s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.141277] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.312s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.141462] env[65385]: INFO nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] During sync_power_state the instance has a pending task (deleting). Skip. [ 1159.141653] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.176490] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1159.177581] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce070f7-098e-41b0-abea-0949a50c194d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.188031] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1159.188031] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a039c0f9-ce94-4f29-8517-a238f2398bf5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.260020] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1159.261050] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1159.261050] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleting the datastore file [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1159.261256] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2825b9c6-6bf0-4681-b4d4-b5c96e5dd5d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.268374] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1159.268374] env[65385]: value = "task-4454561" [ 1159.268374] env[65385]: _type = "Task" [ 1159.268374] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.279078] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454561, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.404815] env[65385]: INFO nova.compute.manager [None req-effdad08-4623-49e2-82e4-2ec084ca0021 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance to original state: 'active' [ 1159.623179] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.623483] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.623740] env[65385]: DEBUG nova.objects.instance [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'resources' on Instance uuid 8db87973-e093-4da1-8534-62ea6896a57d {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.631924] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.652449] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1159.654733] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871190', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'name': 'volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b92bc73a-9666-435e-b387-18df7c838f9b', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'serial': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1159.654733] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3728005-b5bd-4a14-8c54-e48e15fdbf4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.671633] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b20cec5-6365-46a5-9172-cf295b3fd4ba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.698053] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10/volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.698384] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93f1b241-4f68-454a-add8-291757cf866a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.719579] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1159.719579] env[65385]: value = "task-4454562" [ 1159.719579] env[65385]: _type = "Task" [ 1159.719579] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.729496] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454562, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.779071] env[65385]: DEBUG oslo_vmware.api [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454561, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218536} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.779342] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1159.779518] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1159.779691] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1159.804914] env[65385]: INFO nova.scheduler.client.report [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted allocations for instance df419705-e7e8-47b4-b9a0-2f1cf2638f33 [ 1159.826220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.826474] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.826653] env[65385]: DEBUG nova.compute.manager [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Going to confirm migration 9 {{(pid=65385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1160.231580] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454562, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.267080] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d10645d-0d7d-4a47-8e92-c595cd680adf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.274295] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f30816-c02a-4fb2-8fbe-e7e132423b42 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.306129] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d923da-60c7-4877-b23d-123f59eca432 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.309891] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.316656] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9ecb45-a061-4cff-9df0-d74fdebfcede {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.335864] env[65385]: DEBUG nova.compute.provider_tree [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.337405] env[65385]: WARNING neutronclient.v2_0.client [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.399213] env[65385]: WARNING neutronclient.v2_0.client [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.399610] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.399778] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1160.400047] env[65385]: DEBUG nova.network.neutron [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1160.400269] env[65385]: DEBUG nova.objects.instance [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lazy-loading 'info_cache' on Instance uuid 5873d9d9-ff2d-4228-a4ed-b07717ba082e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.477857] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.478169] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "900639b6-9b98-436f-aaad-ea8391f67393" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1160.478399] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "900639b6-9b98-436f-aaad-ea8391f67393-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.478602] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "900639b6-9b98-436f-aaad-ea8391f67393-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1160.478774] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "900639b6-9b98-436f-aaad-ea8391f67393-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1160.480953] env[65385]: INFO nova.compute.manager [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Terminating instance [ 1160.536684] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "6c826914-4c6b-456b-a403-2373d30c803f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.536972] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1160.537233] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "6c826914-4c6b-456b-a403-2373d30c803f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.537442] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1160.537613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1160.540144] env[65385]: INFO nova.compute.manager [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Terminating instance [ 1160.696046] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.696278] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1160.696569] env[65385]: INFO nova.compute.manager [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Attaching volume c9bf37ba-bb9d-492e-98bf-bd450ee85fbc to /dev/sdb [ 1160.732646] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454562, 'name': ReconfigVM_Task, 'duration_secs': 0.957123} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.733015] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10/volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.738965] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d88f1ac-ae02-4da6-82f4-02a22d1d1e72 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.750368] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f961c51-f304-4afc-8aa2-c454a3a30f06 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.759236] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe8116a-5a52-4046-a354-b700439a8ecc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.762340] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1160.762340] env[65385]: value = "task-4454563" [ 1160.762340] env[65385]: _type = "Task" [ 1160.762340] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.771350] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454563, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.775228] env[65385]: DEBUG nova.virt.block_device [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updating existing volume attachment record: 918d3c88-5311-43ce-bb10-e179dc3f8956 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1160.841038] env[65385]: DEBUG nova.scheduler.client.report [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1160.985019] env[65385]: DEBUG nova.compute.manager [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1160.985333] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1160.986298] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798c4728-4e02-4c24-af64-8900f2fa8611 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.994997] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1160.995293] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ef9fc34-d321-4028-b8bf-f5bf06ab8100 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.002099] env[65385]: DEBUG oslo_vmware.api [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1161.002099] env[65385]: value = "task-4454565" [ 1161.002099] env[65385]: _type = "Task" [ 1161.002099] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.011595] env[65385]: DEBUG oslo_vmware.api [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.049656] env[65385]: DEBUG nova.compute.manager [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1161.049656] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1161.049656] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f33376f-6771-4978-8ff0-f3971a27704f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.054960] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.055266] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a72dcad5-291b-4341-9a08-c313bb9b07e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.062790] env[65385]: DEBUG oslo_vmware.api [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1161.062790] env[65385]: value = "task-4454567" [ 1161.062790] env[65385]: _type = "Task" [ 1161.062790] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.074979] env[65385]: DEBUG oslo_vmware.api [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454567, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.113388] env[65385]: DEBUG nova.compute.manager [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1161.113769] env[65385]: DEBUG nova.compute.manager [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing instance network info cache due to event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1161.114028] env[65385]: DEBUG oslo_concurrency.lockutils [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Acquiring lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.114283] env[65385]: DEBUG oslo_concurrency.lockutils [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Acquired lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1161.114450] env[65385]: DEBUG nova.network.neutron [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1161.272962] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454563, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.346085] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.722s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.348611] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.039s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.348850] env[65385]: DEBUG nova.objects.instance [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'resources' on Instance uuid df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.372433] env[65385]: INFO nova.scheduler.client.report [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted allocations for instance 8db87973-e093-4da1-8534-62ea6896a57d [ 1161.406429] env[65385]: WARNING neutronclient.v2_0.client [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.406886] env[65385]: WARNING openstack [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.407257] env[65385]: WARNING openstack [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.498471] env[65385]: WARNING openstack [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.498863] env[65385]: WARNING openstack [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.515121] env[65385]: DEBUG oslo_vmware.api [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454565, 'name': PowerOffVM_Task, 'duration_secs': 0.203062} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.515382] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1161.515539] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1161.515800] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb774ec3-97d3-43eb-b571-5a0d8d063acc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.568767] env[65385]: WARNING neutronclient.v2_0.client [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.569456] env[65385]: WARNING openstack [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.569838] env[65385]: WARNING openstack [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.585752] env[65385]: DEBUG oslo_vmware.api [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454567, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.589820] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1161.590057] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1161.590240] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleting the datastore file [datastore2] 900639b6-9b98-436f-aaad-ea8391f67393 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1161.590523] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a8463e4-3119-423c-a8b5-72ad3aa43a40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.597907] env[65385]: DEBUG oslo_vmware.api [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for the task: (returnval){ [ 1161.597907] env[65385]: value = "task-4454570" [ 1161.597907] env[65385]: _type = "Task" [ 1161.597907] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.607072] env[65385]: DEBUG oslo_vmware.api [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454570, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.616864] env[65385]: WARNING neutronclient.v2_0.client [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.617579] env[65385]: WARNING openstack [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.617948] env[65385]: WARNING openstack [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.705420] env[65385]: DEBUG nova.network.neutron [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance_info_cache with network_info: [{"id": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "address": "fa:16:3e:b0:35:a7", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e4dcc22-4c", "ovs_interfaceid": "7e4dcc22-4cf8-46f3-9acd-c5ad6685f818", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1161.774788] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454563, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.776615] env[65385]: WARNING openstack [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.777086] env[65385]: WARNING openstack [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.851888] env[65385]: DEBUG nova.objects.instance [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'numa_topology' on Instance uuid df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.869317] env[65385]: WARNING neutronclient.v2_0.client [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.869992] env[65385]: WARNING openstack [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.870434] env[65385]: WARNING openstack [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.882807] env[65385]: DEBUG oslo_concurrency.lockutils [None req-be320f5f-f122-4639-a905-3e727aa6f0f8 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "8db87973-e093-4da1-8534-62ea6896a57d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.727s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.979493] env[65385]: DEBUG nova.network.neutron [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updated VIF entry in instance network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1161.979830] env[65385]: DEBUG nova.network.neutron [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapcd731def-a6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1162.075256] env[65385]: DEBUG oslo_vmware.api [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454567, 'name': PowerOffVM_Task, 'duration_secs': 0.764265} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.075539] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1162.075703] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1162.075964] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d9c6b1d-e693-489f-a0cb-f9004086a922 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.109145] env[65385]: DEBUG oslo_vmware.api [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Task: {'id': task-4454570, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118197} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.109424] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1162.109608] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1162.109784] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1162.109955] env[65385]: INFO nova.compute.manager [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1162.110232] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1162.110437] env[65385]: DEBUG nova.compute.manager [-] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1162.110556] env[65385]: DEBUG nova.network.neutron [-] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1162.110792] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.111345] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.111603] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.143227] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1162.143442] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1162.143690] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleting the datastore file [datastore2] 6c826914-4c6b-456b-a403-2373d30c803f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1162.144046] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f569405b-e119-42d7-b678-7a26f78cd99c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.150169] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.154470] env[65385]: DEBUG oslo_vmware.api [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1162.154470] env[65385]: value = "task-4454572" [ 1162.154470] env[65385]: _type = "Task" [ 1162.154470] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.163900] env[65385]: DEBUG oslo_vmware.api [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454572, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.208779] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-5873d9d9-ff2d-4228-a4ed-b07717ba082e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1162.209091] env[65385]: DEBUG nova.objects.instance [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lazy-loading 'migration_context' on Instance uuid 5873d9d9-ff2d-4228-a4ed-b07717ba082e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.276549] env[65385]: DEBUG oslo_vmware.api [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454563, 'name': ReconfigVM_Task, 'duration_secs': 1.077022} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.276880] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871190', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'name': 'volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b92bc73a-9666-435e-b387-18df7c838f9b', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'serial': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1162.355360] env[65385]: DEBUG nova.objects.base [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1162.486989] env[65385]: DEBUG oslo_concurrency.lockutils [req-4e5cb28a-7132-4c1f-aebc-3d07ebb984f6 req-ed38f7a6-ef9e-43af-bbde-365aa8b18e6d service nova] Releasing lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1162.500349] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df893e4-7325-4f0f-97e2-f0fe3bed82b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.510282] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8df2d17-b970-4f81-9815-edae39057264 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.542331] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71243dfb-4ab1-460d-9bfa-2934c4deefd2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.550628] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38273aa5-e7b5-4821-829b-4b3f28c8dc56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.564681] env[65385]: DEBUG nova.compute.provider_tree [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.667819] env[65385]: DEBUG oslo_vmware.api [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454572, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165136} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.667819] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1162.667819] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1162.667819] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1162.667819] env[65385]: INFO nova.compute.manager [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1162.667819] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1162.667819] env[65385]: DEBUG nova.compute.manager [-] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1162.667819] env[65385]: DEBUG nova.network.neutron [-] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1162.667819] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.667819] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.667819] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.712221] env[65385]: DEBUG nova.objects.base [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Object Instance<5873d9d9-ff2d-4228-a4ed-b07717ba082e> lazy-loaded attributes: info_cache,migration_context {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1162.713237] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4751589b-8308-4c74-be88-b8d93f30d64b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.736776] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.738862] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e22e58c2-5949-41e7-a554-983e8dfe1aea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.745705] env[65385]: DEBUG oslo_vmware.api [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1162.745705] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cc3376-955f-9352-614f-15d43e418c7b" [ 1162.745705] env[65385]: _type = "Task" [ 1162.745705] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.757419] env[65385]: DEBUG oslo_vmware.api [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cc3376-955f-9352-614f-15d43e418c7b, 'name': SearchDatastore_Task, 'duration_secs': 0.009005} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.757744] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.861927] env[65385]: DEBUG nova.network.neutron [-] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1162.988454] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.068512] env[65385]: DEBUG nova.scheduler.client.report [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1163.142587] env[65385]: DEBUG nova.compute.manager [req-91cd7079-edf5-4df8-b35b-a8ee337d4108 req-2dfc8601-b360-4b5f-97c4-dd373084e030 service nova] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Received event network-vif-deleted-2ab1f4d8-5eff-4f35-8dd5-a81d06ca7f08 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1163.233362] env[65385]: DEBUG nova.compute.manager [req-dbbf7b9f-2a88-491e-ab85-e78ee8d2acc0 req-07d265d1-bec4-4a56-982f-5b26ee3827b4 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Received event network-vif-deleted-92c45f41-92b4-40c3-8516-30189d29e33d {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1163.233590] env[65385]: INFO nova.compute.manager [req-dbbf7b9f-2a88-491e-ab85-e78ee8d2acc0 req-07d265d1-bec4-4a56-982f-5b26ee3827b4 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Neutron deleted interface 92c45f41-92b4-40c3-8516-30189d29e33d; detaching it from the instance and deleting it from the info cache [ 1163.233768] env[65385]: DEBUG nova.network.neutron [req-dbbf7b9f-2a88-491e-ab85-e78ee8d2acc0 req-07d265d1-bec4-4a56-982f-5b26ee3827b4 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.318178] env[65385]: DEBUG nova.objects.instance [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'flavor' on Instance uuid b92bc73a-9666-435e-b387-18df7c838f9b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.367774] env[65385]: INFO nova.compute.manager [-] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Took 1.25 seconds to deallocate network for instance. [ 1163.476557] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.476808] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.560653] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.574698] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.226s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.577617] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.820s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.711922] env[65385]: DEBUG nova.network.neutron [-] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.736717] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ebfeb1dc-523c-4373-96e5-8423ed59df89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.749205] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b82ba66-cb25-466d-96f1-7a8dcbf5bfd2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.780936] env[65385]: DEBUG nova.compute.manager [req-dbbf7b9f-2a88-491e-ab85-e78ee8d2acc0 req-07d265d1-bec4-4a56-982f-5b26ee3827b4 service nova] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Detach interface failed, port_id=92c45f41-92b4-40c3-8516-30189d29e33d, reason: Instance 6c826914-4c6b-456b-a403-2373d30c803f could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1163.823652] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa751f45-eaf0-4430-adcf-f6a2c1a552b8 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.814s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.825185] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.264s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.874675] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.978953] env[65385]: DEBUG nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1164.086252] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8cad5f1c-5273-4d3d-8e60-4e1f90ddcb3e tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.072s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.087102] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 10.258s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.087283] env[65385]: INFO nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] During sync_power_state the instance has a pending task (shelving_image_uploading). Skip. [ 1164.087452] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.087880] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.100s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.088058] env[65385]: INFO nova.compute.manager [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Unshelving [ 1164.215058] env[65385]: INFO nova.compute.manager [-] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Took 1.55 seconds to deallocate network for instance. [ 1164.241358] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3218d2e-2d41-4fb8-8013-b7eecac0d609 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.249710] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff6f9e7-6b18-4bc5-8bf4-b3c870bfd696 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.283769] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70522a93-63a1-4086-8104-ed3b58f75e27 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.292065] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df90ab1-7248-478d-ac17-0bb3283b5a98 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.307255] env[65385]: DEBUG nova.compute.provider_tree [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.326881] env[65385]: INFO nova.compute.manager [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Detaching volume fd0b7de1-8094-45a0-a553-bc8d5424ff10 [ 1164.361719] env[65385]: INFO nova.virt.block_device [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Attempting to driver detach volume fd0b7de1-8094-45a0-a553-bc8d5424ff10 from mountpoint /dev/sdb [ 1164.362029] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1164.362312] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871190', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'name': 'volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b92bc73a-9666-435e-b387-18df7c838f9b', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'serial': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1164.363285] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64bfa5c-19f5-45e7-b80c-d5fea8c2fd5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.386400] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383ce2d4-618f-4c3e-830e-89de53be79be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.394751] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d443ce0-33e5-44cd-a820-53f10347315f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.416175] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990227a0-daa1-4858-ad8b-dda4a8fab288 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.431986] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] The volume has not been displaced from its original location: [datastore1] volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10/volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1164.437317] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1164.437611] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69349acc-fdae-4675-86b5-572e09eae23a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.457215] env[65385]: DEBUG oslo_vmware.api [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1164.457215] env[65385]: value = "task-4454574" [ 1164.457215] env[65385]: _type = "Task" [ 1164.457215] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.467741] env[65385]: DEBUG oslo_vmware.api [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454574, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.498504] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.724241] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.810541] env[65385]: DEBUG nova.scheduler.client.report [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1164.968047] env[65385]: DEBUG oslo_vmware.api [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454574, 'name': ReconfigVM_Task, 'duration_secs': 0.219013} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.968333] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1164.973012] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c881eaf8-3595-49f6-b468-07263df841b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.990671] env[65385]: DEBUG oslo_vmware.api [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1164.990671] env[65385]: value = "task-4454575" [ 1164.990671] env[65385]: _type = "Task" [ 1164.990671] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.999697] env[65385]: DEBUG oslo_vmware.api [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454575, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.117339] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1165.325193] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1165.325552] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1165.326869] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ceee8d7-659f-49f9-abec-eef1242457eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.348159] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610108dc-610f-435e-ae7c-35d8ed229776 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.374435] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc/volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1165.374758] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0c36977-1ab6-45fb-9ccf-0efeaf5d35a1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.395466] env[65385]: DEBUG oslo_vmware.api [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1165.395466] env[65385]: value = "task-4454576" [ 1165.395466] env[65385]: _type = "Task" [ 1165.395466] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.404718] env[65385]: DEBUG oslo_vmware.api [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454576, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.501303] env[65385]: DEBUG oslo_vmware.api [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454575, 'name': ReconfigVM_Task, 'duration_secs': 0.152454} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.501609] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871190', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'name': 'volume-fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b92bc73a-9666-435e-b387-18df7c838f9b', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10', 'serial': 'fd0b7de1-8094-45a0-a553-bc8d5424ff10'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1165.821942] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.244s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1165.825023] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.950s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1165.825699] env[65385]: DEBUG nova.objects.instance [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lazy-loading 'resources' on Instance uuid 900639b6-9b98-436f-aaad-ea8391f67393 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.905458] env[65385]: DEBUG oslo_vmware.api [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454576, 'name': ReconfigVM_Task, 'duration_secs': 0.379159} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.905729] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc/volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.910527] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2c9f439-4c68-491e-8bb5-b49377285e1c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.926883] env[65385]: DEBUG oslo_vmware.api [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1165.926883] env[65385]: value = "task-4454577" [ 1165.926883] env[65385]: _type = "Task" [ 1165.926883] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.939319] env[65385]: DEBUG oslo_vmware.api [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454577, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.044742] env[65385]: DEBUG nova.objects.instance [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'flavor' on Instance uuid b92bc73a-9666-435e-b387-18df7c838f9b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.390681] env[65385]: INFO nova.scheduler.client.report [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted allocation for migration 63618a0e-63c2-49ca-84d5-466f3f3f0c0d [ 1166.440243] env[65385]: DEBUG oslo_vmware.api [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454577, 'name': ReconfigVM_Task, 'duration_secs': 0.149156} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.440592] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1166.490949] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b091932-7136-4e09-8a43-9620e23a1185 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.499621] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc9335df-56a0-49e5-8db0-c747be110c52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.531043] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9cebd9-76a4-476d-879d-eb72e65737ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.539603] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71bfb89-4054-45a0-add8-25353fa4a53b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.557046] env[65385]: DEBUG nova.compute.provider_tree [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1166.897174] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9ce85486-12bf-4761-8802-3bed5fdc4445 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.070s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.970037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.970037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.970037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.970037] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.970255] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.973187] env[65385]: INFO nova.compute.manager [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Terminating instance [ 1167.061040] env[65385]: DEBUG nova.scheduler.client.report [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1167.064410] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b726a74a-2a2b-4c44-a86d-5993ebbe3788 tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.240s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.477522] env[65385]: DEBUG nova.compute.manager [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1167.477852] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1167.479180] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f959f857-5570-4e76-ac1a-c3a26524eb9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.486306] env[65385]: DEBUG nova.objects.instance [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'flavor' on Instance uuid 62c30ded-c851-477f-b1d4-921268a6ef1c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.492397] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "9de8e3db-35b5-43a8-b677-d815055a0a51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.492830] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.498321] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1167.498525] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31eee735-82ec-4dec-9a40-3e678b712ea7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.509320] env[65385]: DEBUG oslo_vmware.api [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1167.509320] env[65385]: value = "task-4454578" [ 1167.509320] env[65385]: _type = "Task" [ 1167.509320] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.522142] env[65385]: DEBUG oslo_vmware.api [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.566485] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.570129] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.071s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.571632] env[65385]: INFO nova.compute.claims [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1167.598911] env[65385]: INFO nova.scheduler.client.report [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Deleted allocations for instance 900639b6-9b98-436f-aaad-ea8391f67393 [ 1167.807912] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.808209] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.808444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "b92bc73a-9666-435e-b387-18df7c838f9b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.808669] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.808859] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.811276] env[65385]: INFO nova.compute.manager [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Terminating instance [ 1167.995211] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4b634b63-0280-4b0d-b948-e1b39b7c9420 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.299s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.996336] env[65385]: DEBUG nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1168.023897] env[65385]: DEBUG oslo_vmware.api [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454578, 'name': PowerOffVM_Task, 'duration_secs': 0.202183} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.024183] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1168.024317] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1168.024581] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb36f70a-2b0b-4ad2-8f6d-672ecec3b940 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.097353] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1168.097576] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1168.097753] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleting the datastore file [datastore2] 5873d9d9-ff2d-4228-a4ed-b07717ba082e {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1168.098031] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e472ec55-58e9-4504-871e-00c5325869b3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.107614] env[65385]: DEBUG oslo_vmware.api [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1168.107614] env[65385]: value = "task-4454580" [ 1168.107614] env[65385]: _type = "Task" [ 1168.107614] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.108188] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7a1001f9-db81-48ec-aebd-c700b17610a5 tempest-AttachInterfacesTestJSON-1562818112 tempest-AttachInterfacesTestJSON-1562818112-project-member] Lock "900639b6-9b98-436f-aaad-ea8391f67393" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.630s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.120433] env[65385]: DEBUG oslo_vmware.api [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.315431] env[65385]: DEBUG nova.compute.manager [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1168.315666] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1168.316599] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0456097-ad1d-4f33-aac2-3710038a04b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.326565] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1168.326565] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-688d5038-6962-44d6-85b6-e662d956f1b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.334050] env[65385]: DEBUG oslo_vmware.api [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1168.334050] env[65385]: value = "task-4454581" [ 1168.334050] env[65385]: _type = "Task" [ 1168.334050] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.343897] env[65385]: DEBUG oslo_vmware.api [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454581, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.520397] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.621106] env[65385]: DEBUG oslo_vmware.api [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285741} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.624265] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1168.624443] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1168.624662] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1168.624840] env[65385]: INFO nova.compute.manager [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1168.625174] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1168.625677] env[65385]: DEBUG nova.compute.manager [-] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1168.625784] env[65385]: DEBUG nova.network.neutron [-] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1168.626101] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.626720] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.627046] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.703422] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.751357] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15728ec-55aa-4b2d-9f5b-f024d44e65f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.760107] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6f2f73-b4d3-4773-a89b-f306090b0bef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.793304] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c31662-9008-4029-b7fa-ed2a8d783d66 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.802926] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e59bd00-9144-48db-8737-3ebe878e5dbd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.819143] env[65385]: DEBUG nova.compute.provider_tree [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.845063] env[65385]: DEBUG oslo_vmware.api [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454581, 'name': PowerOffVM_Task, 'duration_secs': 0.201493} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.845283] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1168.845592] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1168.845758] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b811fa3-ab52-466f-96ff-8b7f890941c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.908898] env[65385]: INFO nova.compute.manager [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Rebuilding instance [ 1168.936541] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1168.936541] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1168.936541] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleting the datastore file [datastore2] b92bc73a-9666-435e-b387-18df7c838f9b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1168.936541] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cccad29-be55-4de8-8927-e556b6585481 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.947192] env[65385]: DEBUG oslo_vmware.api [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for the task: (returnval){ [ 1168.947192] env[65385]: value = "task-4454583" [ 1168.947192] env[65385]: _type = "Task" [ 1168.947192] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.967500] env[65385]: DEBUG oslo_vmware.api [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454583, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.968626] env[65385]: DEBUG nova.compute.manager [req-ba1a3f39-59ba-4a5f-9a73-a45ba9cfdc52 req-801f892b-5c69-44d2-9d48-5cf6dd04aade service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Received event network-vif-deleted-7e4dcc22-4cf8-46f3-9acd-c5ad6685f818 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1168.968848] env[65385]: INFO nova.compute.manager [req-ba1a3f39-59ba-4a5f-9a73-a45ba9cfdc52 req-801f892b-5c69-44d2-9d48-5cf6dd04aade service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Neutron deleted interface 7e4dcc22-4cf8-46f3-9acd-c5ad6685f818; detaching it from the instance and deleting it from the info cache [ 1168.969062] env[65385]: DEBUG nova.network.neutron [req-ba1a3f39-59ba-4a5f-9a73-a45ba9cfdc52 req-801f892b-5c69-44d2-9d48-5cf6dd04aade service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1168.978078] env[65385]: DEBUG nova.compute.manager [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1168.978911] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab306d9-aabe-4459-9670-d5eb17b1f643 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.322502] env[65385]: DEBUG nova.scheduler.client.report [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1169.439106] env[65385]: DEBUG nova.network.neutron [-] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1169.460042] env[65385]: DEBUG oslo_vmware.api [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Task: {'id': task-4454583, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18587} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.460193] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1169.460473] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1169.460552] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1169.460712] env[65385]: INFO nova.compute.manager [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1169.460984] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1169.461231] env[65385]: DEBUG nova.compute.manager [-] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1169.461347] env[65385]: DEBUG nova.network.neutron [-] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1169.461707] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1169.462288] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.462574] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.473504] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55b048cc-0e3d-47e3-8bba-5fb24727623e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.483958] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a32fe87-9190-41a1-93de-9743d5a6b7ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.499663] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1169.518663] env[65385]: DEBUG nova.compute.manager [req-ba1a3f39-59ba-4a5f-9a73-a45ba9cfdc52 req-801f892b-5c69-44d2-9d48-5cf6dd04aade service nova] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Detach interface failed, port_id=7e4dcc22-4cf8-46f3-9acd-c5ad6685f818, reason: Instance 5873d9d9-ff2d-4228-a4ed-b07717ba082e could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1169.828371] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.829110] env[65385]: DEBUG nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1169.831882] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.108s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.832106] env[65385]: DEBUG nova.objects.instance [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'resources' on Instance uuid 6c826914-4c6b-456b-a403-2373d30c803f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.940744] env[65385]: INFO nova.compute.manager [-] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Took 1.31 seconds to deallocate network for instance. [ 1169.985386] env[65385]: DEBUG nova.compute.manager [req-85376277-6e15-4598-9098-b3ae73f858f4 req-d6276814-45fe-427b-98a7-ebd61c64d8ba service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Received event network-vif-deleted-2341618c-8b69-4d79-8e10-8bec1c24a786 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1169.985585] env[65385]: INFO nova.compute.manager [req-85376277-6e15-4598-9098-b3ae73f858f4 req-d6276814-45fe-427b-98a7-ebd61c64d8ba service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Neutron deleted interface 2341618c-8b69-4d79-8e10-8bec1c24a786; detaching it from the instance and deleting it from the info cache [ 1169.986245] env[65385]: DEBUG nova.network.neutron [req-85376277-6e15-4598-9098-b3ae73f858f4 req-d6276814-45fe-427b-98a7-ebd61c64d8ba service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1169.999553] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.999854] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b1b15f8-435f-45ca-8b96-fa15c1bafcfc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.011954] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1170.011954] env[65385]: value = "task-4454584" [ 1170.011954] env[65385]: _type = "Task" [ 1170.011954] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.029758] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.338123] env[65385]: DEBUG nova.compute.utils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1170.342241] env[65385]: DEBUG nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1170.342470] env[65385]: DEBUG nova.network.neutron [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1170.342787] env[65385]: WARNING neutronclient.v2_0.client [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.343105] env[65385]: WARNING neutronclient.v2_0.client [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.344244] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.344244] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.404460] env[65385]: DEBUG nova.policy [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662871f375e441eb98d814cf7d7ac5c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bc8afc699e34f059f18d8244980bc1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1170.421206] env[65385]: DEBUG nova.network.neutron [-] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.448274] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1170.494662] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-659c4d16-b67c-4140-9aae-9ff11efc6db8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.508527] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92e6514-d37d-461b-b63e-e851d29b7b47 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.530695] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454584, 'name': PowerOffVM_Task, 'duration_secs': 0.241851} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.530892] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1170.535208] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487081c9-6b3c-446f-ae19-2884732cb024 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.548441] env[65385]: DEBUG nova.compute.manager [req-85376277-6e15-4598-9098-b3ae73f858f4 req-d6276814-45fe-427b-98a7-ebd61c64d8ba service nova] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Detach interface failed, port_id=2341618c-8b69-4d79-8e10-8bec1c24a786, reason: Instance b92bc73a-9666-435e-b387-18df7c838f9b could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1170.552969] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5be6bc3-b3e9-459e-a4e4-8d6616760c9b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.584334] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9791c38b-3823-45d2-8c1e-fecbf943f703 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.592770] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1947c20d-3ee7-4cf9-a806-31519f19dbe5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.607304] env[65385]: DEBUG nova.compute.provider_tree [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.627224] env[65385]: INFO nova.compute.manager [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Detaching volume c9bf37ba-bb9d-492e-98bf-bd450ee85fbc [ 1170.671149] env[65385]: INFO nova.virt.block_device [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Attempting to driver detach volume c9bf37ba-bb9d-492e-98bf-bd450ee85fbc from mountpoint /dev/sdb [ 1170.671408] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1170.671598] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1170.672505] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c384a2f9-c5e2-4373-8410-75bf7379b38f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.695197] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2936c4-d70e-48c1-baee-4ea5a8146c4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.702792] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41cfa90-c969-4e20-a2fc-1062fa9fd0d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.726229] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9bfd96-2cb1-4f42-b6aa-4fddfbce20ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.742603] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The volume has not been displaced from its original location: [datastore1] volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc/volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1170.747896] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1170.748228] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e8c874e-f000-4178-b1ae-875558c86b43 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.763008] env[65385]: DEBUG nova.network.neutron [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Successfully created port: 56f375aa-8266-43ae-b0bd-5866243423aa {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1170.767549] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1170.767549] env[65385]: value = "task-4454585" [ 1170.767549] env[65385]: _type = "Task" [ 1170.767549] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.776346] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454585, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.852446] env[65385]: DEBUG nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1170.924053] env[65385]: INFO nova.compute.manager [-] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Took 1.46 seconds to deallocate network for instance. [ 1171.111027] env[65385]: DEBUG nova.scheduler.client.report [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1171.279030] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454585, 'name': ReconfigVM_Task, 'duration_secs': 0.215238} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.279330] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1171.284146] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa9e0fc6-d661-4be4-8e75-e611c69c2d87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.300342] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1171.300342] env[65385]: value = "task-4454586" [ 1171.300342] env[65385]: _type = "Task" [ 1171.300342] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.309393] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454586, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.431192] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.616218] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.619071] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.502s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.619651] env[65385]: DEBUG nova.objects.instance [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'pci_requests' on Instance uuid df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.641066] env[65385]: INFO nova.scheduler.client.report [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted allocations for instance 6c826914-4c6b-456b-a403-2373d30c803f [ 1171.811451] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454586, 'name': ReconfigVM_Task, 'duration_secs': 0.377544} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.811768] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1171.861991] env[65385]: DEBUG nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1171.893403] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1171.893703] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1171.893865] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1171.894062] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1171.894200] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1171.894406] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1171.894730] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1171.894903] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1171.895076] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1171.895235] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1171.895412] env[65385]: DEBUG nova.virt.hardware [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1171.896278] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfc9196-60e2-4847-b5c4-f425e1ce7fbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.904773] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10c28a7-b47e-4b83-95dc-77cbeb8d7708 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.123994] env[65385]: DEBUG nova.objects.instance [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'numa_topology' on Instance uuid df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.148092] env[65385]: DEBUG oslo_concurrency.lockutils [None req-f988110e-65d5-4b7c-a7db-47a5a13f8833 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "6c826914-4c6b-456b-a403-2373d30c803f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.611s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.199508] env[65385]: DEBUG nova.compute.manager [req-d5af775e-d294-48cf-9a67-4c7adf4c011d req-b8b642b8-af34-4faa-b215-5e4ef2d16fd4 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Received event network-vif-plugged-56f375aa-8266-43ae-b0bd-5866243423aa {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1172.199508] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5af775e-d294-48cf-9a67-4c7adf4c011d req-b8b642b8-af34-4faa-b215-5e4ef2d16fd4 service nova] Acquiring lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.199508] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5af775e-d294-48cf-9a67-4c7adf4c011d req-b8b642b8-af34-4faa-b215-5e4ef2d16fd4 service nova] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.200165] env[65385]: DEBUG oslo_concurrency.lockutils [req-d5af775e-d294-48cf-9a67-4c7adf4c011d req-b8b642b8-af34-4faa-b215-5e4ef2d16fd4 service nova] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.200450] env[65385]: DEBUG nova.compute.manager [req-d5af775e-d294-48cf-9a67-4c7adf4c011d req-b8b642b8-af34-4faa-b215-5e4ef2d16fd4 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] No waiting events found dispatching network-vif-plugged-56f375aa-8266-43ae-b0bd-5866243423aa {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1172.201235] env[65385]: WARNING nova.compute.manager [req-d5af775e-d294-48cf-9a67-4c7adf4c011d req-b8b642b8-af34-4faa-b215-5e4ef2d16fd4 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Received unexpected event network-vif-plugged-56f375aa-8266-43ae-b0bd-5866243423aa for instance with vm_state building and task_state spawning. [ 1172.286896] env[65385]: DEBUG nova.network.neutron [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Successfully updated port: 56f375aa-8266-43ae-b0bd-5866243423aa {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1172.627808] env[65385]: INFO nova.compute.claims [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1172.790135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "refresh_cache-e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.790135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "refresh_cache-e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1172.790135] env[65385]: DEBUG nova.network.neutron [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1172.873145] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.873476] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cfe5a38-1ad4-4619-a0a8-2320849d8f7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.883378] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1172.883378] env[65385]: value = "task-4454587" [ 1172.883378] env[65385]: _type = "Task" [ 1172.883378] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.894192] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1172.894884] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1172.894884] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1172.895445] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6f7a10-18d2-49a0-b3b7-14177663ba42 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.916519] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29439b8a-d9c2-49ab-9be9-d0b8456654c5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.924647] env[65385]: WARNING nova.virt.vmwareapi.driver [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1172.924923] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1172.925700] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570f2e66-ffbb-4308-8324-d36282bb09c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.933986] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1172.934267] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39b22d73-caf3-4381-92d8-f131ad70ada6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.006872] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1173.007106] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1173.007284] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleting the datastore file [datastore1] 62c30ded-c851-477f-b1d4-921268a6ef1c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1173.007550] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e405b971-3162-4bf6-a297-031efcafc6f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.015964] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1173.015964] env[65385]: value = "task-4454589" [ 1173.015964] env[65385]: _type = "Task" [ 1173.015964] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.026076] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.292940] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.293414] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.332866] env[65385]: DEBUG nova.network.neutron [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1173.353626] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.354182] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.434880] env[65385]: WARNING neutronclient.v2_0.client [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.435452] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.435882] env[65385]: WARNING openstack [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.454534] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.454793] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.517893] env[65385]: DEBUG nova.network.neutron [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Updating instance_info_cache with network_info: [{"id": "56f375aa-8266-43ae-b0bd-5866243423aa", "address": "fa:16:3e:f7:98:99", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56f375aa-82", "ovs_interfaceid": "56f375aa-8266-43ae-b0bd-5866243423aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1173.530342] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454589, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158873} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.530655] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.530808] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.531021] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.784890] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84013623-b074-4117-a49d-71d1dde9920b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.793444] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71285982-ed0c-46df-a95f-9e7506fdbe43 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.824325] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f6a2be-9351-491a-a03c-2a2f342e0690 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.832875] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353253ff-fe94-41df-82eb-317fef0096a9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.846839] env[65385]: DEBUG nova.compute.provider_tree [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.957346] env[65385]: DEBUG nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1174.024624] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "refresh_cache-e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1174.025063] env[65385]: DEBUG nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Instance network_info: |[{"id": "56f375aa-8266-43ae-b0bd-5866243423aa", "address": "fa:16:3e:f7:98:99", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56f375aa-82", "ovs_interfaceid": "56f375aa-8266-43ae-b0bd-5866243423aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1174.025522] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:98:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56f375aa-8266-43ae-b0bd-5866243423aa', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1174.033456] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1174.034083] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1174.034407] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-432c9f1e-52f0-4d67-9a1a-5ed7199aedce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.052418] env[65385]: INFO nova.virt.block_device [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Booting with volume c9bf37ba-bb9d-492e-98bf-bd450ee85fbc at /dev/sdb [ 1174.062054] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1174.062054] env[65385]: value = "task-4454590" [ 1174.062054] env[65385]: _type = "Task" [ 1174.062054] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.071647] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454590, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.091046] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-079d0c41-4e7e-4712-8915-447ab3efd2dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.101254] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b73db76-9fa8-4659-8c1f-f7162a48562b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.134744] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ffa5c475-7b9f-4bd2-94fa-00ea6d7f3eba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.144821] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b855fa-4d5e-4577-8ccd-a3d570c991fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.174552] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d208f05-a92a-405d-b508-d3e7f9b78466 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.181972] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172f4675-9da5-4e72-b04f-cc76435f3c74 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.195952] env[65385]: DEBUG nova.virt.block_device [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updating existing volume attachment record: 37cce358-ad1e-4508-bcc0-428527d9184a {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1174.237956] env[65385]: DEBUG nova.compute.manager [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Received event network-changed-56f375aa-8266-43ae-b0bd-5866243423aa {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1174.237956] env[65385]: DEBUG nova.compute.manager [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Refreshing instance network info cache due to event network-changed-56f375aa-8266-43ae-b0bd-5866243423aa. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1174.238119] env[65385]: DEBUG oslo_concurrency.lockutils [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Acquiring lock "refresh_cache-e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.238217] env[65385]: DEBUG oslo_concurrency.lockutils [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Acquired lock "refresh_cache-e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1174.238372] env[65385]: DEBUG nova.network.neutron [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Refreshing network info cache for port 56f375aa-8266-43ae-b0bd-5866243423aa {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1174.350155] env[65385]: DEBUG nova.scheduler.client.report [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1174.476793] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.573191] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454590, 'name': CreateVM_Task, 'duration_secs': 0.341683} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.573367] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1174.573932] env[65385]: WARNING neutronclient.v2_0.client [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.574312] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.574461] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1174.574792] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1174.575062] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbcb49b5-8d0b-4e59-803d-0db370484b9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.580429] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1174.580429] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bccf8-fb92-3812-1e49-e72cbbc47872" [ 1174.580429] env[65385]: _type = "Task" [ 1174.580429] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.591107] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bccf8-fb92-3812-1e49-e72cbbc47872, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.740714] env[65385]: WARNING neutronclient.v2_0.client [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.742094] env[65385]: WARNING openstack [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.742488] env[65385]: WARNING openstack [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.835979] env[65385]: WARNING openstack [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.836444] env[65385]: WARNING openstack [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.854799] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.236s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.855172] env[65385]: WARNING neutronclient.v2_0.client [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.857513] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.337s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.859022] env[65385]: INFO nova.compute.claims [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1174.902767] env[65385]: INFO nova.network.neutron [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating port cd731def-a6f7-40f9-b6ad-a10d2fd72984 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1174.906337] env[65385]: WARNING neutronclient.v2_0.client [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.906609] env[65385]: WARNING openstack [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.906957] env[65385]: WARNING openstack [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.992741] env[65385]: DEBUG nova.network.neutron [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Updated VIF entry in instance network info cache for port 56f375aa-8266-43ae-b0bd-5866243423aa. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1174.993176] env[65385]: DEBUG nova.network.neutron [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Updating instance_info_cache with network_info: [{"id": "56f375aa-8266-43ae-b0bd-5866243423aa", "address": "fa:16:3e:f7:98:99", "network": {"id": "5e2b3694-eda2-4f0a-880e-90b1d7efcefe", "bridge": "br-int", "label": "tempest-ServersTestJSON-279085577-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bc8afc699e34f059f18d8244980bc1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56f375aa-82", "ovs_interfaceid": "56f375aa-8266-43ae-b0bd-5866243423aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1175.092052] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527bccf8-fb92-3812-1e49-e72cbbc47872, 'name': SearchDatastore_Task, 'duration_secs': 0.011532} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.093396] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1175.093680] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1175.093929] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.094171] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.094360] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1175.094917] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29ede79f-0484-4249-be1d-1293ea85a85a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.104199] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1175.104379] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1175.105205] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ce2d4a1-8263-4ee2-9ebd-22c2aafd6bf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.111665] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1175.111665] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9c338-ba78-5c31-96c1-615830e7b579" [ 1175.111665] env[65385]: _type = "Task" [ 1175.111665] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.119851] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9c338-ba78-5c31-96c1-615830e7b579, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.496117] env[65385]: DEBUG oslo_concurrency.lockutils [req-ea8f99e5-6d39-4b39-a989-e4eb43ad36f4 req-b5c64682-42be-4646-be20-849573bb9e78 service nova] Releasing lock "refresh_cache-e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1175.624825] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a9c338-ba78-5c31-96c1-615830e7b579, 'name': SearchDatastore_Task, 'duration_secs': 0.00982} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.626026] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2763df2-73e6-4511-8047-a93129f171ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.632415] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1175.632415] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e24452-4015-1c9e-b5f4-1f6e36530254" [ 1175.632415] env[65385]: _type = "Task" [ 1175.632415] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.641451] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e24452-4015-1c9e-b5f4-1f6e36530254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.007792] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b88c84b2-e7d1-49f1-9563-3881f87e01d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.015655] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ae8f64-5c57-4785-a733-f33795272327 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.048120] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa9b1a0-b3c1-4b82-a8e0-0e49801659dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.056602] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4a9e13-ba3a-43f7-825e-d24cb04ca7ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.070874] env[65385]: DEBUG nova.compute.provider_tree [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.144203] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e24452-4015-1c9e-b5f4-1f6e36530254, 'name': SearchDatastore_Task, 'duration_secs': 0.010023} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.144404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1176.144681] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f/e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1176.144928] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-901bf494-069e-4229-beee-13137dfd7202 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.152541] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1176.152541] env[65385]: value = "task-4454591" [ 1176.152541] env[65385]: _type = "Task" [ 1176.152541] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.160936] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454591, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.326363] env[65385]: DEBUG nova.compute.manager [req-e5d49540-7729-4346-bc3b-a2968205f8bf req-71e8fb08-fcdf-45ca-8716-ce7b400b62d7 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-vif-plugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1176.326363] env[65385]: DEBUG oslo_concurrency.lockutils [req-e5d49540-7729-4346-bc3b-a2968205f8bf req-71e8fb08-fcdf-45ca-8716-ce7b400b62d7 service nova] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.326363] env[65385]: DEBUG oslo_concurrency.lockutils [req-e5d49540-7729-4346-bc3b-a2968205f8bf req-71e8fb08-fcdf-45ca-8716-ce7b400b62d7 service nova] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.326363] env[65385]: DEBUG oslo_concurrency.lockutils [req-e5d49540-7729-4346-bc3b-a2968205f8bf req-71e8fb08-fcdf-45ca-8716-ce7b400b62d7 service nova] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.326742] env[65385]: DEBUG nova.compute.manager [req-e5d49540-7729-4346-bc3b-a2968205f8bf req-71e8fb08-fcdf-45ca-8716-ce7b400b62d7 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] No waiting events found dispatching network-vif-plugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1176.326897] env[65385]: WARNING nova.compute.manager [req-e5d49540-7729-4346-bc3b-a2968205f8bf req-71e8fb08-fcdf-45ca-8716-ce7b400b62d7 service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received unexpected event network-vif-plugged-cd731def-a6f7-40f9-b6ad-a10d2fd72984 for instance with vm_state shelved_offloaded and task_state spawning. [ 1176.329669] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1176.329669] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1176.329882] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1176.329964] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1176.330082] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1176.330227] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1176.330432] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1176.330585] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1176.330749] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1176.330878] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1176.331130] env[65385]: DEBUG nova.virt.hardware [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1176.332422] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af8a149-6a6a-4e43-9e01-3557e2eaf8e0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.342114] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35783565-2a00-4cfc-8f59-c5807b5963a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.358263] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:92:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5e55898-473b-4cdf-9d34-0b704e923915', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1176.366086] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1176.366425] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1176.366671] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ed3b91b-b31d-422b-99ff-bba04964f2b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.388918] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1176.388918] env[65385]: value = "task-4454592" [ 1176.388918] env[65385]: _type = "Task" [ 1176.388918] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.398564] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454592, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.430043] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.430297] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.430490] env[65385]: DEBUG nova.network.neutron [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1176.574397] env[65385]: DEBUG nova.scheduler.client.report [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1176.668078] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454591, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462696} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.668372] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f/e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1176.668641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1176.668910] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-636edd6d-4a7e-447d-a40f-fdafff3e397c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.676889] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1176.676889] env[65385]: value = "task-4454593" [ 1176.676889] env[65385]: _type = "Task" [ 1176.676889] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.687667] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454593, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.899049] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454592, 'name': CreateVM_Task, 'duration_secs': 0.395099} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.899341] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1176.899877] env[65385]: WARNING neutronclient.v2_0.client [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1176.900269] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.900632] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.901114] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1176.901384] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf3b2a32-4d02-42c7-bab4-e4019b86fe31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.906825] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1176.906825] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cfc179-0283-8c5c-28d7-36cbe03cc4c3" [ 1176.906825] env[65385]: _type = "Task" [ 1176.906825] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.915467] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cfc179-0283-8c5c-28d7-36cbe03cc4c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.934110] env[65385]: WARNING neutronclient.v2_0.client [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.935203] env[65385]: WARNING openstack [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.935203] env[65385]: WARNING openstack [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.060167] env[65385]: WARNING openstack [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.060578] env[65385]: WARNING openstack [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.080460] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.081012] env[65385]: DEBUG nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1177.083809] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.636s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.084081] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.086701] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.656s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.086921] env[65385]: DEBUG nova.objects.instance [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lazy-loading 'resources' on Instance uuid b92bc73a-9666-435e-b387-18df7c838f9b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.112505] env[65385]: INFO nova.scheduler.client.report [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted allocations for instance 5873d9d9-ff2d-4228-a4ed-b07717ba082e [ 1177.127756] env[65385]: WARNING neutronclient.v2_0.client [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1177.128427] env[65385]: WARNING openstack [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.128849] env[65385]: WARNING openstack [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.186807] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454593, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087282} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.187072] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1177.187848] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100382f3-35d3-46a5-b97e-93a7b4a512f1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.210946] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f/e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1177.213302] env[65385]: DEBUG nova.network.neutron [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd731def-a6", "ovs_interfaceid": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1177.214541] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d04f9724-eae9-459c-86c6-bfadbb7bdcba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.229847] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.238387] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1177.238387] env[65385]: value = "task-4454594" [ 1177.238387] env[65385]: _type = "Task" [ 1177.238387] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.247739] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.257849] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6597c7ec46194a73199de3cb3c6f3694',container_format='bare',created_at=2025-11-14T16:56:17Z,direct_url=,disk_format='vmdk',id=1b100802-b12e-4f1f-a998-6080dbb6821d,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-308151370-shelved',owner='b27c092a0fd345ea953956469c21f8bb',properties=ImageMetaProps,protected=,size=31662080,status='active',tags=,updated_at=2025-11-14T16:56:34Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1177.258091] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1177.258245] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1177.258418] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1177.258557] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1177.258698] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1177.258903] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1177.259068] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1177.259232] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1177.259388] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1177.259556] env[65385]: DEBUG nova.virt.hardware [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1177.260402] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676b6017-2d03-4387-b84d-928a5522468d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.268587] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6bce7b-b75d-4ac6-9dfd-fe7a06ebd9df {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.283629] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:32:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4a8fd90-153b-494f-b76a-299eb05c03f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd731def-a6f7-40f9-b6ad-a10d2fd72984', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1177.290972] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1177.291255] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1177.291560] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74c898ef-6d7a-41e1-a1ad-5f58314139a5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.314603] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1177.314603] env[65385]: value = "task-4454595" [ 1177.314603] env[65385]: _type = "Task" [ 1177.314603] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.324373] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454595, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.417486] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cfc179-0283-8c5c-28d7-36cbe03cc4c3, 'name': SearchDatastore_Task, 'duration_secs': 0.011642} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.417791] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.418034] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1177.418296] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.418444] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1177.418617] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1177.418954] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eeeb8fb9-e8ee-45df-b150-0b5c3d5c77c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.427806] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1177.428051] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1177.428866] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-761c39ac-a217-4a3d-a413-b843de14db53 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.435279] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1177.435279] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ed082f-219e-e5f3-8343-228defb94484" [ 1177.435279] env[65385]: _type = "Task" [ 1177.435279] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.444708] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ed082f-219e-e5f3-8343-228defb94484, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.594844] env[65385]: DEBUG nova.compute.utils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1177.596387] env[65385]: DEBUG nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1177.596990] env[65385]: DEBUG nova.network.neutron [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1177.596990] env[65385]: WARNING neutronclient.v2_0.client [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1177.597348] env[65385]: WARNING neutronclient.v2_0.client [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1177.598380] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1177.598489] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1177.622194] env[65385]: DEBUG oslo_concurrency.lockutils [None req-191d7899-0d23-40cb-8584-27aab1ea1d41 tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "5873d9d9-ff2d-4228-a4ed-b07717ba082e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.652s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.657086] env[65385]: DEBUG nova.policy [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c1a9d4194964403a8d11abfad4c65a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94cc5bc221c0455ea760f5022db6bdc4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1177.729880] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b085aad-6ebb-43ca-824a-fad11f35af9e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.738133] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acc5618-7832-492f-b85d-1c62eca853c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.750071] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454594, 'name': ReconfigVM_Task, 'duration_secs': 0.282249} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.773870] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Reconfigured VM instance instance-00000073 to attach disk [datastore2] e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f/e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1177.775207] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3517943-273f-4a71-9e29-3eccea6eb710 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.777353] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84a7a21-0b0a-4927-8fef-042781ec7ac6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.786900] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092b371d-0303-4b16-a34e-44365d184608 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.791098] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1177.791098] env[65385]: value = "task-4454596" [ 1177.791098] env[65385]: _type = "Task" [ 1177.791098] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.804252] env[65385]: DEBUG nova.compute.provider_tree [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.812102] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454596, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.824856] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454595, 'name': CreateVM_Task, 'duration_secs': 0.376077} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.825025] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1177.825570] env[65385]: WARNING neutronclient.v2_0.client [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1177.825930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.826105] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1177.826852] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1177.826852] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ff6466d-50ce-473f-92a8-544b4128b56b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.832124] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1177.832124] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b2e478-fbc0-4281-f99c-c6b099f92864" [ 1177.832124] env[65385]: _type = "Task" [ 1177.832124] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.842380] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b2e478-fbc0-4281-f99c-c6b099f92864, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.949180] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ed082f-219e-e5f3-8343-228defb94484, 'name': SearchDatastore_Task, 'duration_secs': 0.010472} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.950752] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63f34b93-9f30-4ee9-bfec-13ef55fc4b6c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.959456] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1177.959456] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5209d0a9-6ed7-5bfc-1e3a-25240dc09a1a" [ 1177.959456] env[65385]: _type = "Task" [ 1177.959456] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.975430] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5209d0a9-6ed7-5bfc-1e3a-25240dc09a1a, 'name': SearchDatastore_Task, 'duration_secs': 0.011098} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.975851] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.976265] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1177.976665] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9be45794-9701-4b7f-be15-350d44a6c6db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.980748] env[65385]: DEBUG nova.network.neutron [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Successfully created port: b4e12be3-8c2c-49d7-8f67-781a1ee6680e {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1177.991455] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1177.991455] env[65385]: value = "task-4454597" [ 1177.991455] env[65385]: _type = "Task" [ 1177.991455] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.002026] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454597, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.106797] env[65385]: DEBUG nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1178.304298] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454596, 'name': Rename_Task, 'duration_secs': 0.159559} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.304617] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1178.304961] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0ed33a2-07b9-441b-ba3a-5ea940710bcd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.307630] env[65385]: DEBUG nova.scheduler.client.report [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1178.319774] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1178.319774] env[65385]: value = "task-4454598" [ 1178.319774] env[65385]: _type = "Task" [ 1178.319774] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.329765] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454598, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.348170] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.348511] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Processing image 1b100802-b12e-4f1f-a998-6080dbb6821d {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.348786] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d/1b100802-b12e-4f1f-a998-6080dbb6821d.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.348960] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d/1b100802-b12e-4f1f-a998-6080dbb6821d.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.349336] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.349680] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95af13cd-893b-4dfe-a749-0eed488b60d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.360349] env[65385]: DEBUG nova.compute.manager [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1178.360611] env[65385]: DEBUG nova.compute.manager [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing instance network info cache due to event network-changed-cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1178.360852] env[65385]: DEBUG oslo_concurrency.lockutils [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Acquiring lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.361130] env[65385]: DEBUG oslo_concurrency.lockutils [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Acquired lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.361279] env[65385]: DEBUG nova.network.neutron [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Refreshing network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1178.365209] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.365466] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1178.366497] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b30d555d-1fa1-4802-b5b7-e4b61b293952 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.375794] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1178.375794] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ac933a-526f-1386-df48-32b32a3dce76" [ 1178.375794] env[65385]: _type = "Task" [ 1178.375794] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.388110] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ac933a-526f-1386-df48-32b32a3dce76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.502237] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454597, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504371} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.502541] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1178.502761] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1178.503054] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0c01a86-8da9-41b6-be63-5c2f4add0192 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.510870] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1178.510870] env[65385]: value = "task-4454599" [ 1178.510870] env[65385]: _type = "Task" [ 1178.510870] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.521253] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454599, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.813608] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.816378] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.340s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.818156] env[65385]: INFO nova.compute.claims [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1178.831575] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454598, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.845899] env[65385]: INFO nova.scheduler.client.report [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Deleted allocations for instance b92bc73a-9666-435e-b387-18df7c838f9b [ 1178.868527] env[65385]: WARNING neutronclient.v2_0.client [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1178.868527] env[65385]: WARNING openstack [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1178.868786] env[65385]: WARNING openstack [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1178.888191] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Preparing fetch location {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1178.888641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Fetch image to [datastore1] OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757/OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757.vmdk {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1178.888746] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Downloading stream optimized image 1b100802-b12e-4f1f-a998-6080dbb6821d to [datastore1] OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757/OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757.vmdk on the data store datastore1 as vApp {{(pid=65385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1178.888914] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Downloading image file data 1b100802-b12e-4f1f-a998-6080dbb6821d to the ESX as VM named 'OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757' {{(pid=65385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1179.022802] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454599, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073654} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.023112] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1179.023959] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ea42e4-d9ed-4daa-933d-d626e393590d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.049173] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1179.050310] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1179.050310] env[65385]: value = "resgroup-9" [ 1179.050310] env[65385]: _type = "ResourcePool" [ 1179.050310] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1179.050585] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5abd7593-456d-46bd-8a23-c5e16a30621d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.065601] env[65385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7c49aa03-c989-4639-8190-91d70b20a7ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.090813] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease: (returnval){ [ 1179.090813] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263dde9-430d-a7e2-ed72-d4b755157b27" [ 1179.090813] env[65385]: _type = "HttpNfcLease" [ 1179.090813] env[65385]: } obtained for vApp import into resource pool (val){ [ 1179.090813] env[65385]: value = "resgroup-9" [ 1179.090813] env[65385]: _type = "ResourcePool" [ 1179.090813] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1179.091136] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the lease: (returnval){ [ 1179.091136] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263dde9-430d-a7e2-ed72-d4b755157b27" [ 1179.091136] env[65385]: _type = "HttpNfcLease" [ 1179.091136] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1179.092578] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1179.092578] env[65385]: value = "task-4454600" [ 1179.092578] env[65385]: _type = "Task" [ 1179.092578] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.103028] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1179.103028] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263dde9-430d-a7e2-ed72-d4b755157b27" [ 1179.103028] env[65385]: _type = "HttpNfcLease" [ 1179.103028] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1179.106428] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454600, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.116541] env[65385]: DEBUG nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1179.122100] env[65385]: WARNING openstack [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.122519] env[65385]: WARNING openstack [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.144182] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1179.144476] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1179.144634] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1179.144919] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1179.145103] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1179.145346] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1179.145583] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1179.145740] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1179.145938] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1179.146122] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1179.146294] env[65385]: DEBUG nova.virt.hardware [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1179.147246] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2d451e-50f5-487b-94ed-0f4fbac12d91 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.156827] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12181b58-d1dd-49b8-92aa-8b3645b56856 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.223387] env[65385]: WARNING neutronclient.v2_0.client [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.224147] env[65385]: WARNING openstack [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.224518] env[65385]: WARNING openstack [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.341116] env[65385]: DEBUG oslo_vmware.api [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454598, 'name': PowerOnVM_Task, 'duration_secs': 0.572823} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.342165] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1179.342432] env[65385]: INFO nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1179.342615] env[65385]: DEBUG nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1179.343533] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd744996-1dbc-42bb-8453-501e424b1941 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.348541] env[65385]: DEBUG nova.network.neutron [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updated VIF entry in instance network info cache for port cd731def-a6f7-40f9-b6ad-a10d2fd72984. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1179.348907] env[65385]: DEBUG nova.network.neutron [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [{"id": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "address": "fa:16:3e:85:32:59", "network": {"id": "39efc686-bcc7-4971-b412-a3b58c1f1e18", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1551320485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b27c092a0fd345ea953956469c21f8bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4a8fd90-153b-494f-b76a-299eb05c03f9", "external-id": "nsx-vlan-transportzone-981", "segmentation_id": 981, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd731def-a6", "ovs_interfaceid": "cd731def-a6f7-40f9-b6ad-a10d2fd72984", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1179.354594] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e16d8dbe-aa44-4a17-bbaf-32c3a24cf39d tempest-AttachVolumeNegativeTest-1972219862 tempest-AttachVolumeNegativeTest-1972219862-project-member] Lock "b92bc73a-9666-435e-b387-18df7c838f9b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.546s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.416576] env[65385]: DEBUG nova.compute.manager [req-8850bfc3-ae4a-4826-aaf9-06f747579f07 req-335313dc-e340-411c-9da4-ea250a7de5fa service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Received event network-vif-plugged-b4e12be3-8c2c-49d7-8f67-781a1ee6680e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1179.416894] env[65385]: DEBUG oslo_concurrency.lockutils [req-8850bfc3-ae4a-4826-aaf9-06f747579f07 req-335313dc-e340-411c-9da4-ea250a7de5fa service nova] Acquiring lock "9de8e3db-35b5-43a8-b677-d815055a0a51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1179.417246] env[65385]: DEBUG oslo_concurrency.lockutils [req-8850bfc3-ae4a-4826-aaf9-06f747579f07 req-335313dc-e340-411c-9da4-ea250a7de5fa service nova] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1179.417491] env[65385]: DEBUG oslo_concurrency.lockutils [req-8850bfc3-ae4a-4826-aaf9-06f747579f07 req-335313dc-e340-411c-9da4-ea250a7de5fa service nova] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.417676] env[65385]: DEBUG nova.compute.manager [req-8850bfc3-ae4a-4826-aaf9-06f747579f07 req-335313dc-e340-411c-9da4-ea250a7de5fa service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] No waiting events found dispatching network-vif-plugged-b4e12be3-8c2c-49d7-8f67-781a1ee6680e {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1179.417842] env[65385]: WARNING nova.compute.manager [req-8850bfc3-ae4a-4826-aaf9-06f747579f07 req-335313dc-e340-411c-9da4-ea250a7de5fa service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Received unexpected event network-vif-plugged-b4e12be3-8c2c-49d7-8f67-781a1ee6680e for instance with vm_state building and task_state spawning. [ 1179.519751] env[65385]: DEBUG nova.network.neutron [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Successfully updated port: b4e12be3-8c2c-49d7-8f67-781a1ee6680e {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1179.604669] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1179.604669] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263dde9-430d-a7e2-ed72-d4b755157b27" [ 1179.604669] env[65385]: _type = "HttpNfcLease" [ 1179.604669] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1179.607849] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454600, 'name': ReconfigVM_Task, 'duration_secs': 0.278497} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.608127] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 62c30ded-c851-477f-b1d4-921268a6ef1c/62c30ded-c851-477f-b1d4-921268a6ef1c.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1179.609245] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'boot_index': 0, 'encryption_secret_uuid': None, 'device_type': 'disk', 'guest_format': None, 'encryption_options': None, 'encrypted': False, 'encryption_format': None, 'device_name': '/dev/sda', 'disk_bus': None, 'size': 0, 'image_id': 'c3e9f50a-a10e-4952-ab03-d986996039fa'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'}, 'device_type': None, 'guest_format': None, 'attachment_id': '37cce358-ad1e-4508-bcc0-428527d9184a', 'mount_device': '/dev/sdb', 'disk_bus': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=65385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1179.609440] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1179.609618] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1179.610378] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2813b7-0860-48ea-a89d-3f6838bc37fd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.630368] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e731afd-7869-46c1-9dac-9b608b8b4ca0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.656372] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc/volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1179.656686] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c02c14e-f95b-4b47-9e92-caefeee92395 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.676327] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1179.676327] env[65385]: value = "task-4454602" [ 1179.676327] env[65385]: _type = "Task" [ 1179.676327] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.685504] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454602, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.858369] env[65385]: DEBUG oslo_concurrency.lockutils [req-acedb912-f254-4fb9-b9d4-9b23b7e064bc req-b3ca11ae-8f4d-4a09-abd4-29a0c0b0caff service nova] Releasing lock "refresh_cache-df419705-e7e8-47b4-b9a0-2f1cf2638f33" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1179.867401] env[65385]: INFO nova.compute.manager [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Took 15.38 seconds to build instance. [ 1179.960959] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0315985f-646b-478c-83c8-2b1c1da3a86d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.974522] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea86ec8c-5b50-4be3-82ce-d796452b5347 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.021626] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb4a198-a4f3-43a6-b70c-ac3b56e99485 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.026090] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "refresh_cache-9de8e3db-35b5-43a8-b677-d815055a0a51" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.026355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "refresh_cache-9de8e3db-35b5-43a8-b677-d815055a0a51" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1180.026610] env[65385]: DEBUG nova.network.neutron [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1180.035667] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cf83a3-43cb-429c-b4d3-96fe2e8394e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.057456] env[65385]: DEBUG nova.compute.provider_tree [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.102746] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1180.102746] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263dde9-430d-a7e2-ed72-d4b755157b27" [ 1180.102746] env[65385]: _type = "HttpNfcLease" [ 1180.102746] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1180.187721] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454602, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.369971] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7b0e3f6c-3b91-4374-82d3-932d8842561e tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.893s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.405746] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.406088] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.530165] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1180.530600] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1180.562746] env[65385]: DEBUG nova.scheduler.client.report [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1180.569353] env[65385]: DEBUG nova.network.neutron [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1180.590040] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1180.590493] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1180.606790] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1180.606790] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263dde9-430d-a7e2-ed72-d4b755157b27" [ 1180.606790] env[65385]: _type = "HttpNfcLease" [ 1180.606790] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1180.606939] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1180.606939] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5263dde9-430d-a7e2-ed72-d4b755157b27" [ 1180.606939] env[65385]: _type = "HttpNfcLease" [ 1180.606939] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1180.607726] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04728fd3-c4b1-444f-9a22-1bc98e7119fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.616888] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52745889-d1b6-b937-0e9a-53a9a7c4a7bd/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1180.617145] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating HTTP connection to write to file with size = 31662080 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52745889-d1b6-b937-0e9a-53a9a7c4a7bd/disk-0.vmdk. {{(pid=65385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1180.686744] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b5eeb29e-32c2-4d3b-9370-4c51a5313016 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.695757] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454602, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.722278] env[65385]: WARNING neutronclient.v2_0.client [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1180.722941] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1180.723318] env[65385]: WARNING openstack [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1180.809326] env[65385]: DEBUG nova.network.neutron [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Updating instance_info_cache with network_info: [{"id": "b4e12be3-8c2c-49d7-8f67-781a1ee6680e", "address": "fa:16:3e:45:39:cb", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4e12be3-8c", "ovs_interfaceid": "b4e12be3-8c2c-49d7-8f67-781a1ee6680e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1180.838717] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.843607] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.843758] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1181.069811] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.070394] env[65385]: DEBUG nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1181.190554] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454602, 'name': ReconfigVM_Task, 'duration_secs': 1.320829} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.190934] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc/volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1181.195701] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22a74a39-d90c-4e7d-9efb-d1a8ac4d1d6d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.212756] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1181.212756] env[65385]: value = "task-4454604" [ 1181.212756] env[65385]: _type = "Task" [ 1181.212756] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.222500] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.313874] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "refresh_cache-9de8e3db-35b5-43a8-b677-d815055a0a51" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1181.314353] env[65385]: DEBUG nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Instance network_info: |[{"id": "b4e12be3-8c2c-49d7-8f67-781a1ee6680e", "address": "fa:16:3e:45:39:cb", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4e12be3-8c", "ovs_interfaceid": "b4e12be3-8c2c-49d7-8f67-781a1ee6680e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1181.314863] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.315120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.315306] env[65385]: DEBUG nova.compute.manager [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1181.315730] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:39:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4e12be3-8c2c-49d7-8f67-781a1ee6680e', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.323361] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1181.327012] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a813b1f-7fdc-4959-be7c-b165b9df6626 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.329949] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1181.330576] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0f9a22b-3633-4db5-a148-c295401619b7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.354288] env[65385]: DEBUG nova.compute.manager [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1181.354943] env[65385]: DEBUG nova.objects.instance [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'flavor' on Instance uuid e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.358922] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1181.358922] env[65385]: value = "task-4454605" [ 1181.358922] env[65385]: _type = "Task" [ 1181.358922] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.371089] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454605, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.479953] env[65385]: DEBUG nova.compute.manager [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Received event network-changed-b4e12be3-8c2c-49d7-8f67-781a1ee6680e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1181.480184] env[65385]: DEBUG nova.compute.manager [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Refreshing instance network info cache due to event network-changed-b4e12be3-8c2c-49d7-8f67-781a1ee6680e. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1181.480421] env[65385]: DEBUG oslo_concurrency.lockutils [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Acquiring lock "refresh_cache-9de8e3db-35b5-43a8-b677-d815055a0a51" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.480640] env[65385]: DEBUG oslo_concurrency.lockutils [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Acquired lock "refresh_cache-9de8e3db-35b5-43a8-b677-d815055a0a51" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1181.480765] env[65385]: DEBUG nova.network.neutron [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Refreshing network info cache for port b4e12be3-8c2c-49d7-8f67-781a1ee6680e {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1181.575918] env[65385]: DEBUG nova.compute.utils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1181.579442] env[65385]: DEBUG nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1181.579673] env[65385]: DEBUG nova.network.neutron [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1181.580174] env[65385]: WARNING neutronclient.v2_0.client [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.580540] env[65385]: WARNING neutronclient.v2_0.client [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.581343] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.581774] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.595817] env[65385]: DEBUG nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1181.666122] env[65385]: DEBUG nova.policy [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2a287457352470794887b229ea90cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60de0d1c162342209795a180391661f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1181.728371] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454604, 'name': ReconfigVM_Task, 'duration_secs': 0.17064} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.730779] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1181.731503] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93deedbd-e95d-4992-83f2-d670be093f9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.739029] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1181.739029] env[65385]: value = "task-4454606" [ 1181.739029] env[65385]: _type = "Task" [ 1181.739029] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.751239] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454606, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.844461] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.878650] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454605, 'name': CreateVM_Task, 'duration_secs': 0.372322} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.880894] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1181.881557] env[65385]: WARNING neutronclient.v2_0.client [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.882022] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.882253] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1181.882631] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1181.883215] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8aa6d138-66da-4138-af86-1ce5b2f685e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.889606] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1181.889606] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290deef-fe7e-0d1d-eba4-df6173c0a849" [ 1181.889606] env[65385]: _type = "Task" [ 1181.889606] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.900601] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290deef-fe7e-0d1d-eba4-df6173c0a849, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.983839] env[65385]: WARNING neutronclient.v2_0.client [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.984662] env[65385]: WARNING openstack [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.985098] env[65385]: WARNING openstack [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.059507] env[65385]: DEBUG nova.network.neutron [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Successfully created port: 34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1182.182948] env[65385]: WARNING openstack [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.183689] env[65385]: WARNING openstack [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.252028] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454606, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.275780] env[65385]: WARNING neutronclient.v2_0.client [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1182.276489] env[65385]: WARNING openstack [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.276864] env[65385]: WARNING openstack [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.336419] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Completed reading data from the image iterator. {{(pid=65385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1182.336475] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52745889-d1b6-b937-0e9a-53a9a7c4a7bd/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1182.337415] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38023861-c1da-412f-bc6d-d4202cf092ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.345574] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52745889-d1b6-b937-0e9a-53a9a7c4a7bd/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1182.345716] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52745889-d1b6-b937-0e9a-53a9a7c4a7bd/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1182.346035] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ad741aa2-6f81-4f91-87d9-5b6f5e92391f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.348796] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.350380] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.350380] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.350380] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1182.350380] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e526f21e-70b7-4cb7-8ef0-c31d09a45a39 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.360550] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6ee193-40e5-4266-b3c8-6bf87fb3d0cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.369494] env[65385]: DEBUG nova.network.neutron [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Updated VIF entry in instance network info cache for port b4e12be3-8c2c-49d7-8f67-781a1ee6680e. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1182.369857] env[65385]: DEBUG nova.network.neutron [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Updating instance_info_cache with network_info: [{"id": "b4e12be3-8c2c-49d7-8f67-781a1ee6680e", "address": "fa:16:3e:45:39:cb", "network": {"id": "74c61a47-6c24-4b01-a1a7-031d71ac33bf", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2081957845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94cc5bc221c0455ea760f5022db6bdc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4e12be3-8c", "ovs_interfaceid": "b4e12be3-8c2c-49d7-8f67-781a1ee6680e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1182.371407] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1182.380631] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab1ab553-0ded-4eef-ab6a-9abef849e37d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.384554] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3171fc-d3b8-47e5-9757-f9ba3082c438 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.399128] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa887ba1-7d6b-4215-9f87-367c433559e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.403078] env[65385]: DEBUG oslo_vmware.api [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1182.403078] env[65385]: value = "task-4454607" [ 1182.403078] env[65385]: _type = "Task" [ 1182.403078] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.439855] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290deef-fe7e-0d1d-eba4-df6173c0a849, 'name': SearchDatastore_Task, 'duration_secs': 0.01135} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.440230] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.440441] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.440619] env[65385]: INFO nova.compute.manager [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Shelving [ 1182.442090] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178976MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1182.442223] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.442417] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.444791] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1182.445056] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1182.445305] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.445462] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1182.445647] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.446098] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8aee6ce4-121c-46ed-a9d5-4a5e9c3fb7fc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.452434] env[65385]: DEBUG oslo_vmware.api [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.462436] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.462648] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1182.463505] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7b9b789-7fa3-46d7-b879-ec168bbccf7c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.471170] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1182.471170] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52080a99-7eaf-8ad7-c15c-d2851cf708ba" [ 1182.471170] env[65385]: _type = "Task" [ 1182.471170] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.480922] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52080a99-7eaf-8ad7-c15c-d2851cf708ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.607118] env[65385]: DEBUG nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1182.638550] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1182.638830] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1182.639021] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1182.640202] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1182.640202] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1182.640202] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1182.640202] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1182.640202] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1182.640202] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1182.640668] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1182.640668] env[65385]: DEBUG nova.virt.hardware [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1182.641437] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f11d60-568c-426b-bba2-9a3b3907670f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.650948] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfe9d0b-671a-4d2c-bb6d-097faddb0395 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.668842] env[65385]: DEBUG oslo_vmware.rw_handles [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52745889-d1b6-b937-0e9a-53a9a7c4a7bd/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1182.669106] env[65385]: INFO nova.virt.vmwareapi.images [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Downloaded image file data 1b100802-b12e-4f1f-a998-6080dbb6821d [ 1182.669872] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349a0fda-f014-42bd-9d13-726899fb882a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.686622] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe4e034f-ebac-4a89-b84d-7364e9f1f731 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.729640] env[65385]: INFO nova.virt.vmwareapi.images [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] The imported VM was unregistered [ 1182.733955] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Caching image {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1182.734246] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Creating directory with path [datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.734550] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65d5c599-a909-417d-99f4-102b0270b9af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.747563] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Created directory with path [datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.747748] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757/OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757.vmdk to [datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d/1b100802-b12e-4f1f-a998-6080dbb6821d.vmdk. {{(pid=65385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1182.750965] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-43916cfd-6ed1-4fd0-b4e0-de99a4cc2479 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.753232] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454606, 'name': Rename_Task, 'duration_secs': 0.753714} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.753900] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1182.754486] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74450651-0110-4566-91c6-c03d43e58bc4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.759929] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1182.759929] env[65385]: value = "task-4454609" [ 1182.759929] env[65385]: _type = "Task" [ 1182.759929] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.761226] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1182.761226] env[65385]: value = "task-4454610" [ 1182.761226] env[65385]: _type = "Task" [ 1182.761226] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.774109] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454609, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.776803] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454610, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.884126] env[65385]: DEBUG oslo_concurrency.lockutils [req-ab8810b5-71b2-4621-89f6-4d9711454d72 req-8df01377-8d01-4f6b-8cce-6e1615b119c6 service nova] Releasing lock "refresh_cache-9de8e3db-35b5-43a8-b677-d815055a0a51" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1182.914338] env[65385]: DEBUG oslo_vmware.api [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454607, 'name': PowerOffVM_Task, 'duration_secs': 0.26436} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.914766] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1182.915034] env[65385]: DEBUG nova.compute.manager [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1182.916026] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9ccb96-7713-48c7-bae5-104e26030704 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.986058] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52080a99-7eaf-8ad7-c15c-d2851cf708ba, 'name': SearchDatastore_Task, 'duration_secs': 0.017296} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.986913] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d145589e-4f73-4696-9aa3-dcd569def282 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.994378] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1182.994378] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a00bbc-b2e8-7735-dde0-6b87ce449bac" [ 1182.994378] env[65385]: _type = "Task" [ 1182.994378] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.004291] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a00bbc-b2e8-7735-dde0-6b87ce449bac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.276827] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454609, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.280814] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454610, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.431687] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e553630f-6ea0-4821-b0fe-8cfbf208187b tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.116s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.458901] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1183.459242] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7213fcf7-2c22-4e79-8121-2fb3eeac9445 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.475089] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1183.475089] env[65385]: value = "task-4454611" [ 1183.475089] env[65385]: _type = "Task" [ 1183.475089] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b4bca2c8-fc8f-471a-899b-aac648fee9fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 992920c3-af4f-489e-b21f-9d52c24399de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 62c30ded-c851-477f-b1d4-921268a6ef1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 20933973-82ae-498c-b016-3a82bb5dc165 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance df419705-e7e8-47b4-b9a0-2f1cf2638f33 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 9de8e3db-35b5-43a8-b677-d815055a0a51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance fa02564b-c1a8-4f8f-876c-806618e37f5b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1183.486913] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1183.487556] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=100GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '8', 'num_vm_active': '5', 'num_task_None': '2', 'num_os_type_None': '8', 'num_proj_8d45ff3446a84ef2a1945f7a66b747e8': '2', 'io_workload': '2', 'num_proj_6bc8afc699e34f059f18d8244980bc1d': '2', 'num_vm_shelved_offloaded': '1', 'num_task_spawning': '3', 'num_proj_b27c092a0fd345ea953956469c21f8bb': '1', 'num_task_rebuild_spawning': '1', 'num_task_shelving': '1', 'num_proj_1598307c304d46a884a88b3eb5f5d7dc': '1', 'num_task_powering-off': '1', 'num_vm_building': '2', 'num_proj_94cc5bc221c0455ea760f5022db6bdc4': '1', 'num_proj_60de0d1c162342209795a180391661f6': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1183.505491] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.508088] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1183.514209] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a00bbc-b2e8-7735-dde0-6b87ce449bac, 'name': SearchDatastore_Task, 'duration_secs': 0.05632} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.514505] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.514820] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 9de8e3db-35b5-43a8-b677-d815055a0a51/9de8e3db-35b5-43a8-b677-d815055a0a51.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1183.515133] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1a9d908-c4b7-4fed-809b-b17d1b7b006e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.523651] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1183.523651] env[65385]: value = "task-4454613" [ 1183.523651] env[65385]: _type = "Task" [ 1183.523651] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.528882] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1183.529100] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1183.539394] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.545492] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1183.567992] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1183.702115] env[65385]: DEBUG nova.compute.manager [req-dfc6df61-a0f3-4266-ac38-af1d9617f31f req-a4496fc8-c786-49b5-89c0-7f202e7116e3 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Received event network-vif-plugged-34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1183.702339] env[65385]: DEBUG oslo_concurrency.lockutils [req-dfc6df61-a0f3-4266-ac38-af1d9617f31f req-a4496fc8-c786-49b5-89c0-7f202e7116e3 service nova] Acquiring lock "fa02564b-c1a8-4f8f-876c-806618e37f5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.702538] env[65385]: DEBUG oslo_concurrency.lockutils [req-dfc6df61-a0f3-4266-ac38-af1d9617f31f req-a4496fc8-c786-49b5-89c0-7f202e7116e3 service nova] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.702717] env[65385]: DEBUG oslo_concurrency.lockutils [req-dfc6df61-a0f3-4266-ac38-af1d9617f31f req-a4496fc8-c786-49b5-89c0-7f202e7116e3 service nova] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.703976] env[65385]: DEBUG nova.compute.manager [req-dfc6df61-a0f3-4266-ac38-af1d9617f31f req-a4496fc8-c786-49b5-89c0-7f202e7116e3 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] No waiting events found dispatching network-vif-plugged-34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1183.703976] env[65385]: WARNING nova.compute.manager [req-dfc6df61-a0f3-4266-ac38-af1d9617f31f req-a4496fc8-c786-49b5-89c0-7f202e7116e3 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Received unexpected event network-vif-plugged-34d33bf4-6698-4d35-a2ee-636a6e0fdbec for instance with vm_state building and task_state spawning. [ 1183.705048] env[65385]: DEBUG nova.network.neutron [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Successfully updated port: 34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1183.716797] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b574f3ca-6abf-494c-960f-9c0cd2274eb3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.728917] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d3bc5a-4153-4a62-928d-a8e9b4744f81 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.775710] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afee98c-8004-4ddd-a314-194aef2f550d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.785399] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454609, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.791430] env[65385]: DEBUG oslo_vmware.api [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454610, 'name': PowerOnVM_Task, 'duration_secs': 0.52036} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.791924] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1183.792162] env[65385]: DEBUG nova.compute.manager [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1183.793496] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733090b9-75ab-4fd8-b86e-339bea8943aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.799511] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a907fed-6aee-4232-806f-217bf9edb0c0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.816246] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1183.988106] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.037048] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.208481] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.208838] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1184.209073] env[65385]: DEBUG nova.network.neutron [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1184.280840] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454609, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.324650] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1184.334415] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.389759] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.390089] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.390387] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.390639] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.390885] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.396133] env[65385]: INFO nova.compute.manager [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Terminating instance [ 1184.489043] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454611, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.538532] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.712919] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.713546] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.763321] env[65385]: DEBUG nova.network.neutron [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1184.782749] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454609, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.786273] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.786968] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.833295] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1184.833527] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.391s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.833923] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.500s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.834197] env[65385]: DEBUG nova.objects.instance [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1184.862123] env[65385]: WARNING neutronclient.v2_0.client [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1184.862815] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.863187] env[65385]: WARNING openstack [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.902989] env[65385]: DEBUG nova.compute.manager [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1184.903278] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1184.904507] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6f9321-f274-44bc-b8d2-fa20eb0b2e73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.915711] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1184.916068] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e8b90b4-f3e6-42cb-8d28-5caeb78bb7ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.990824] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454611, 'name': PowerOffVM_Task, 'duration_secs': 1.024509} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.991118] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1184.992536] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5294f69b-9807-40a5-b099-60d1aed0aa34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.020491] env[65385]: DEBUG nova.network.neutron [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updating instance_info_cache with network_info: [{"id": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "address": "fa:16:3e:83:87:18", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d33bf4-66", "ovs_interfaceid": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1185.022774] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3392929e-4dbb-44d7-b54f-f089501ae028 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.028171] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1185.028400] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1185.028540] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleting the datastore file [datastore2] e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.032925] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-482ae0cf-a4b4-4128-99fd-852d648ae7de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.051369] env[65385]: DEBUG oslo_vmware.api [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1185.051369] env[65385]: value = "task-4454615" [ 1185.051369] env[65385]: _type = "Task" [ 1185.051369] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.055161] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.068856] env[65385]: DEBUG oslo_vmware.api [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454615, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.280947] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454609, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.527860] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1185.528365] env[65385]: DEBUG nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Instance network_info: |[{"id": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "address": "fa:16:3e:83:87:18", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d33bf4-66", "ovs_interfaceid": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1185.529243] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:87:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34d33bf4-6698-4d35-a2ee-636a6e0fdbec', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1185.539280] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1185.543611] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1185.544891] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Creating Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1185.545182] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5d332da-278b-4e26-951f-97b6c9166888 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.564037] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a68514a0-5861-499a-80de-eb2a1b996192 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.577159] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454613, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.579318] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1185.579318] env[65385]: value = "task-4454616" [ 1185.579318] env[65385]: _type = "Task" [ 1185.579318] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.589255] env[65385]: DEBUG oslo_vmware.api [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454615, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.589255] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1185.589255] env[65385]: value = "task-4454617" [ 1185.589255] env[65385]: _type = "Task" [ 1185.589255] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.596211] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454616, 'name': CreateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.603672] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454617, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.739920] env[65385]: DEBUG nova.compute.manager [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Received event network-changed-34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1185.740341] env[65385]: DEBUG nova.compute.manager [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Refreshing instance network info cache due to event network-changed-34d33bf4-6698-4d35-a2ee-636a6e0fdbec. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1185.740663] env[65385]: DEBUG oslo_concurrency.lockutils [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Acquiring lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.740928] env[65385]: DEBUG oslo_concurrency.lockutils [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Acquired lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.741223] env[65385]: DEBUG nova.network.neutron [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Refreshing network info cache for port 34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1185.780768] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454609, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.863293} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.781095] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757/OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757.vmdk to [datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d/1b100802-b12e-4f1f-a998-6080dbb6821d.vmdk. [ 1185.781304] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Cleaning up location [datastore1] OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1185.781490] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_0ec08de6-cef9-4d31-8f14-16515c6fe757 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.781772] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d19cc70-ffe2-4557-b744-9a3ec7539539 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.789491] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1185.789491] env[65385]: value = "task-4454618" [ 1185.789491] env[65385]: _type = "Task" [ 1185.789491] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.798889] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.837818] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.838118] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.838281] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.845781] env[65385]: DEBUG oslo_concurrency.lockutils [None req-167e1285-b952-4729-bc29-6fd630ae81f5 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.040567] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454613, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.223726} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.040870] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 9de8e3db-35b5-43a8-b677-d815055a0a51/9de8e3db-35b5-43a8-b677-d815055a0a51.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1186.041103] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1186.041387] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c279cc27-7d91-423c-a83e-4bcee7b9083f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.050583] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1186.050583] env[65385]: value = "task-4454619" [ 1186.050583] env[65385]: _type = "Task" [ 1186.050583] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.062446] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.081152] env[65385]: DEBUG oslo_vmware.api [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454615, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.950661} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.085235] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.085515] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1186.085757] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.086018] env[65385]: INFO nova.compute.manager [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1186.086328] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1186.086591] env[65385]: DEBUG nova.compute.manager [-] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1186.086719] env[65385]: DEBUG nova.network.neutron [-] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1186.087014] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.087603] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.087888] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.102450] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454616, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.108885] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454617, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.129532] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.245188] env[65385]: WARNING neutronclient.v2_0.client [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.245913] env[65385]: WARNING openstack [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.246449] env[65385]: WARNING openstack [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.301610] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251289} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.302396] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.302396] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d/1b100802-b12e-4f1f-a998-6080dbb6821d.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1186.302396] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d/1b100802-b12e-4f1f-a998-6080dbb6821d.vmdk to [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1186.302595] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-546e55fe-5071-49d7-b940-f8e9a0f66907 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.312151] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1186.312151] env[65385]: value = "task-4454620" [ 1186.312151] env[65385]: _type = "Task" [ 1186.312151] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.325587] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.396958] env[65385]: WARNING openstack [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.396958] env[65385]: WARNING openstack [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.465492] env[65385]: WARNING neutronclient.v2_0.client [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.466461] env[65385]: WARNING openstack [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.466907] env[65385]: WARNING openstack [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.562046] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.253727} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.564439] env[65385]: DEBUG nova.network.neutron [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updated VIF entry in instance network info cache for port 34d33bf4-6698-4d35-a2ee-636a6e0fdbec. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1186.564859] env[65385]: DEBUG nova.network.neutron [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updating instance_info_cache with network_info: [{"id": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "address": "fa:16:3e:83:87:18", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d33bf4-66", "ovs_interfaceid": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1186.566209] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1186.567291] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc06a6c0-33f3-4a78-b000-f0b4016e0218 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.595604] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 9de8e3db-35b5-43a8-b677-d815055a0a51/9de8e3db-35b5-43a8-b677-d815055a0a51.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.600307] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b96594fd-3459-45a6-9d45-e8c183f1a803 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.632614] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454616, 'name': CreateVM_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.638815] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454617, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.639323] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1186.639323] env[65385]: value = "task-4454621" [ 1186.639323] env[65385]: _type = "Task" [ 1186.639323] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.650496] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454621, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.824334] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454620, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.874163] env[65385]: DEBUG nova.network.neutron [-] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1187.068387] env[65385]: DEBUG oslo_concurrency.lockutils [req-34aa38c9-1496-4fdb-a228-f3d21ffc6ab3 req-a3a1c1b4-680f-4d25-9476-c4ea728257f2 service nova] Releasing lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.097227] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454616, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.112036] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454617, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.152090] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454621, 'name': ReconfigVM_Task, 'duration_secs': 0.439136} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.152920] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 9de8e3db-35b5-43a8-b677-d815055a0a51/9de8e3db-35b5-43a8-b677-d815055a0a51.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1187.153187] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6078e7c7-cb55-4366-8d0b-879fad9100d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.162498] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1187.162498] env[65385]: value = "task-4454623" [ 1187.162498] env[65385]: _type = "Task" [ 1187.162498] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.174646] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454623, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.325329] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454620, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.377473] env[65385]: INFO nova.compute.manager [-] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Took 1.29 seconds to deallocate network for instance. [ 1187.597541] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454616, 'name': CreateVM_Task, 'duration_secs': 1.726246} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.597812] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1187.598432] env[65385]: WARNING neutronclient.v2_0.client [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.607886] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.608097] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1187.608508] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1187.608810] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11470966-db0e-4d99-b489-4b4c5d72e55d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.614422] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454617, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.619138] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1187.619138] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52476c14-2634-7332-e5c6-0c37a9e32e5c" [ 1187.619138] env[65385]: _type = "Task" [ 1187.619138] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.628915] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52476c14-2634-7332-e5c6-0c37a9e32e5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.675698] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454623, 'name': Rename_Task, 'duration_secs': 0.241988} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.676078] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1187.676382] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-029fb8e0-9c43-4af8-becf-bd7ba53bc291 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.685724] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1187.685724] env[65385]: value = "task-4454624" [ 1187.685724] env[65385]: _type = "Task" [ 1187.685724] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.695462] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.768899] env[65385]: DEBUG nova.compute.manager [req-31c9b642-b581-4efb-aeb8-3fe1729f0c0d req-e249fdbd-9bce-427f-91db-a33fdd2451a9 service nova] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Received event network-vif-deleted-56f375aa-8266-43ae-b0bd-5866243423aa {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1187.826291] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454620, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.886851] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.887438] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.887761] env[65385]: DEBUG nova.objects.instance [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'resources' on Instance uuid e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.113257] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454617, 'name': CreateSnapshot_Task, 'duration_secs': 2.173425} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.113509] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Created Snapshot of the VM instance {{(pid=65385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1188.114462] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737d714f-10bc-47e6-ad33-eef77c880757 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.135473] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52476c14-2634-7332-e5c6-0c37a9e32e5c, 'name': SearchDatastore_Task, 'duration_secs': 0.091212} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.135782] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1188.136016] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1188.136266] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.136397] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1188.136563] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1188.136956] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21d10bcb-de3f-4459-a9d6-6be1040d7b5c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.151877] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1188.152263] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1188.152908] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-786d7e81-e89a-46a7-80bf-9bae7c7da7e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.160173] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1188.160173] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c8ebe-4469-56f2-ce18-ec54e0087333" [ 1188.160173] env[65385]: _type = "Task" [ 1188.160173] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.170313] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c8ebe-4469-56f2-ce18-ec54e0087333, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.197323] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454624, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.327162] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454620, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.523159] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cffb59-0c50-4e9c-a914-40591fb1d8d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.531820] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86d2eb1-a37a-424c-b85d-2006bd5f2284 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.566852] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3df4c55-d757-4e8d-89d5-29be2ced2a52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.576017] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0862a0-e6b8-4996-b78d-0a8e62763010 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.593331] env[65385]: DEBUG nova.compute.provider_tree [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.639114] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Creating linked-clone VM from snapshot {{(pid=65385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1188.639842] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dc38bb01-d881-4cc6-960a-782d39145a07 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.650266] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1188.650266] env[65385]: value = "task-4454625" [ 1188.650266] env[65385]: _type = "Task" [ 1188.650266] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.660148] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454625, 'name': CloneVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.671565] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c8ebe-4469-56f2-ce18-ec54e0087333, 'name': SearchDatastore_Task, 'duration_secs': 0.093923} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.672826] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7b2c838-fbe9-4e4e-974e-1e406006e8bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.679331] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1188.679331] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5269e658-5bd0-dbf8-e8f8-cf2f43b7416f" [ 1188.679331] env[65385]: _type = "Task" [ 1188.679331] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.689118] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5269e658-5bd0-dbf8-e8f8-cf2f43b7416f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.701103] env[65385]: DEBUG oslo_vmware.api [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454624, 'name': PowerOnVM_Task, 'duration_secs': 0.645978} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.701560] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1188.701904] env[65385]: INFO nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Took 9.59 seconds to spawn the instance on the hypervisor. [ 1188.702232] env[65385]: DEBUG nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1188.703530] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babc89fa-8f58-4152-8f8e-59079ae7d2cb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.828608] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454620, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.097046] env[65385]: DEBUG nova.scheduler.client.report [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1189.162024] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454625, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.190070] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5269e658-5bd0-dbf8-e8f8-cf2f43b7416f, 'name': SearchDatastore_Task, 'duration_secs': 0.104959} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.190246] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1189.190379] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] fa02564b-c1a8-4f8f-876c-806618e37f5b/fa02564b-c1a8-4f8f-876c-806618e37f5b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1189.190654] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8863bcf6-71ca-4bb1-9b7e-c904733bc7d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.197919] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1189.197919] env[65385]: value = "task-4454626" [ 1189.197919] env[65385]: _type = "Task" [ 1189.197919] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.208387] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.225214] env[65385]: INFO nova.compute.manager [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Took 20.72 seconds to build instance. [ 1189.329150] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454620, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.610806} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.329617] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1b100802-b12e-4f1f-a998-6080dbb6821d/1b100802-b12e-4f1f-a998-6080dbb6821d.vmdk to [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1189.330337] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f43a9a-f09c-46d7-bacd-9cd88980ded9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.354491] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1189.354882] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6f602ef-cc38-4371-b948-13adfd894b8f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.376246] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1189.376246] env[65385]: value = "task-4454627" [ 1189.376246] env[65385]: _type = "Task" [ 1189.376246] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.386080] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454627, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.601650] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.669877] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454625, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.669877] env[65385]: INFO nova.scheduler.client.report [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted allocations for instance e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f [ 1189.712467] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454626, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.727428] env[65385]: DEBUG oslo_concurrency.lockutils [None req-7aedb105-7bd6-49e7-88ec-5aaf0728022e tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.234s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.888072] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.162837] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454625, 'name': CloneVM_Task} progress is 94%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.178535] env[65385]: DEBUG oslo_concurrency.lockutils [None req-874d6116-7e14-4adb-beb5-23b4e476a5e2 tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.788s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.209619] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656786} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.209885] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] fa02564b-c1a8-4f8f-876c-806618e37f5b/fa02564b-c1a8-4f8f-876c-806618e37f5b.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1190.210110] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1190.210367] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94a1ff22-810c-4180-9659-7343afc39658 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.218013] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1190.218013] env[65385]: value = "task-4454628" [ 1190.218013] env[65385]: _type = "Task" [ 1190.218013] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.229215] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454628, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.390970] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454627, 'name': ReconfigVM_Task, 'duration_secs': 0.695659} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.391336] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Reconfigured VM instance instance-0000006a to attach disk [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33/df419705-e7e8-47b4-b9a0-2f1cf2638f33.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1190.391989] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0a28d41-6b48-4379-b28b-05d4b31cba87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.399671] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1190.399671] env[65385]: value = "task-4454629" [ 1190.399671] env[65385]: _type = "Task" [ 1190.399671] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.410970] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454629, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.591675] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "9de8e3db-35b5-43a8-b677-d815055a0a51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.591968] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1190.592464] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "9de8e3db-35b5-43a8-b677-d815055a0a51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.592464] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1190.592582] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.598670] env[65385]: INFO nova.compute.manager [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Terminating instance [ 1190.663424] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454625, 'name': CloneVM_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.731451] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454628, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.253425} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.731451] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1190.731706] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2134434d-b6ab-4023-a641-cee59f9154ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.755319] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] fa02564b-c1a8-4f8f-876c-806618e37f5b/fa02564b-c1a8-4f8f-876c-806618e37f5b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1190.755675] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b173808-5e51-4d7a-8222-63d790138c18 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.776825] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1190.776825] env[65385]: value = "task-4454630" [ 1190.776825] env[65385]: _type = "Task" [ 1190.776825] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.788092] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454630, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.910301] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454629, 'name': Rename_Task, 'duration_secs': 0.334314} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.910604] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1190.910894] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b18f6f1d-d0f6-476a-8885-a51001fd50aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.918759] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1190.918759] env[65385]: value = "task-4454631" [ 1190.918759] env[65385]: _type = "Task" [ 1190.918759] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.928430] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.103540] env[65385]: DEBUG nova.compute.manager [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1191.103869] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1191.104832] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3018d2e9-c022-4732-bd77-0bf5c328ef7b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.114649] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1191.114923] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7af3f05-7800-4f07-8ebc-8a35c0802b02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.122568] env[65385]: DEBUG oslo_vmware.api [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1191.122568] env[65385]: value = "task-4454632" [ 1191.122568] env[65385]: _type = "Task" [ 1191.122568] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.131951] env[65385]: DEBUG oslo_vmware.api [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.167163] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454625, 'name': CloneVM_Task, 'duration_secs': 2.08643} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.167515] env[65385]: INFO nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Created linked-clone VM from snapshot [ 1191.168421] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55da5e97-d4bf-44e5-9413-b0925719636e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.178410] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Uploading image d71dbee8-5170-4c28-88aa-44822c744198 {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1191.212881] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1191.212881] env[65385]: value = "vm-871200" [ 1191.212881] env[65385]: _type = "VirtualMachine" [ 1191.212881] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1191.213287] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-319ff660-1694-46ce-81ea-1b54222e7561 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.222198] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease: (returnval){ [ 1191.222198] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eee9b9-1ed2-c2d5-bc87-008d08134983" [ 1191.222198] env[65385]: _type = "HttpNfcLease" [ 1191.222198] env[65385]: } obtained for exporting VM: (result){ [ 1191.222198] env[65385]: value = "vm-871200" [ 1191.222198] env[65385]: _type = "VirtualMachine" [ 1191.222198] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1191.222535] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the lease: (returnval){ [ 1191.222535] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eee9b9-1ed2-c2d5-bc87-008d08134983" [ 1191.222535] env[65385]: _type = "HttpNfcLease" [ 1191.222535] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1191.232871] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1191.232871] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eee9b9-1ed2-c2d5-bc87-008d08134983" [ 1191.232871] env[65385]: _type = "HttpNfcLease" [ 1191.232871] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1191.287381] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454630, 'name': ReconfigVM_Task, 'duration_secs': 0.312639} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.287596] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Reconfigured VM instance instance-00000075 to attach disk [datastore1] fa02564b-c1a8-4f8f-876c-806618e37f5b/fa02564b-c1a8-4f8f-876c-806618e37f5b.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1191.288993] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f71bfa9-a642-43ae-91a7-4985cfbda8ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.297579] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1191.297579] env[65385]: value = "task-4454634" [ 1191.297579] env[65385]: _type = "Task" [ 1191.297579] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.309715] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454634, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.430471] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454631, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.562451] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "992920c3-af4f-489e-b21f-9d52c24399de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.562738] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "992920c3-af4f-489e-b21f-9d52c24399de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.563035] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "992920c3-af4f-489e-b21f-9d52c24399de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.563315] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "992920c3-af4f-489e-b21f-9d52c24399de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.563662] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "992920c3-af4f-489e-b21f-9d52c24399de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.566354] env[65385]: INFO nova.compute.manager [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Terminating instance [ 1191.633595] env[65385]: DEBUG oslo_vmware.api [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454632, 'name': PowerOffVM_Task, 'duration_secs': 0.238434} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.633896] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1191.634077] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1191.634348] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90a4ae49-765e-4c45-bf3d-5f275b00f815 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.719377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1191.719377] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1191.719644] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleting the datastore file [datastore1] 9de8e3db-35b5-43a8-b677-d815055a0a51 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1191.719912] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0cb89a6-f6c5-4324-9ec2-0d570114bdd5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.732293] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1191.732293] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eee9b9-1ed2-c2d5-bc87-008d08134983" [ 1191.732293] env[65385]: _type = "HttpNfcLease" [ 1191.732293] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1191.734124] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1191.734124] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eee9b9-1ed2-c2d5-bc87-008d08134983" [ 1191.734124] env[65385]: _type = "HttpNfcLease" [ 1191.734124] env[65385]: }. {{(pid=65385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1191.734444] env[65385]: DEBUG oslo_vmware.api [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for the task: (returnval){ [ 1191.734444] env[65385]: value = "task-4454636" [ 1191.734444] env[65385]: _type = "Task" [ 1191.734444] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.735286] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c3cff5-d0b8-4199-8eb6-f89c1908d8bc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.748235] env[65385]: DEBUG oslo_vmware.api [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454636, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.752567] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f7a834-9943-062d-32e0-7a84503af1ab/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1191.752567] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f7a834-9943-062d-32e0-7a84503af1ab/disk-0.vmdk for reading. {{(pid=65385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1191.824546] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454634, 'name': Rename_Task, 'duration_secs': 0.15753} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.824546] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1191.824546] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95b53f6d-31f6-4018-a9c5-23eaac70a862 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.833557] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1191.833557] env[65385]: value = "task-4454637" [ 1191.833557] env[65385]: _type = "Task" [ 1191.833557] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.845373] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454637, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.854279] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1f2dbde3-9e6e-4acb-9a7b-3d01d4dc8056 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.930986] env[65385]: DEBUG oslo_vmware.api [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454631, 'name': PowerOnVM_Task, 'duration_secs': 0.674047} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.931294] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1192.071959] env[65385]: DEBUG nova.compute.manager [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1192.072226] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1192.073746] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed37a879-3824-4778-979e-da222b9976b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.077578] env[65385]: DEBUG nova.compute.manager [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1192.078408] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eefa76c-cacc-4741-809f-6b0b2450fb58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.091451] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1192.091943] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65ab2013-68e0-4285-b059-948b1d54a624 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.099698] env[65385]: DEBUG oslo_vmware.api [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1192.099698] env[65385]: value = "task-4454638" [ 1192.099698] env[65385]: _type = "Task" [ 1192.099698] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.109779] env[65385]: DEBUG oslo_vmware.api [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.250267] env[65385]: DEBUG oslo_vmware.api [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Task: {'id': task-4454636, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173364} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.250365] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1192.250571] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1192.250835] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1192.251019] env[65385]: INFO nova.compute.manager [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1192.251274] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1192.251711] env[65385]: DEBUG nova.compute.manager [-] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1192.251991] env[65385]: DEBUG nova.network.neutron [-] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1192.252226] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.252814] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.254393] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.318298] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.348546] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454637, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.605926] env[65385]: DEBUG oslo_concurrency.lockutils [None req-3505a9be-f940-4889-a983-6bf7fef9970b tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 28.518s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1192.618345] env[65385]: DEBUG oslo_vmware.api [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454638, 'name': PowerOffVM_Task, 'duration_secs': 0.296434} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.618345] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1192.618345] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1192.618678] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3d89d81-380e-411e-892c-9780d93341fb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.642677] env[65385]: DEBUG nova.compute.manager [req-75ee0a07-f207-436f-a46c-e1c0f30613e4 req-f3790d3f-29e2-4bbf-bf04-8acd1f9caf61 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Received event network-vif-deleted-b4e12be3-8c2c-49d7-8f67-781a1ee6680e {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1192.643065] env[65385]: INFO nova.compute.manager [req-75ee0a07-f207-436f-a46c-e1c0f30613e4 req-f3790d3f-29e2-4bbf-bf04-8acd1f9caf61 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Neutron deleted interface b4e12be3-8c2c-49d7-8f67-781a1ee6680e; detaching it from the instance and deleting it from the info cache [ 1192.643359] env[65385]: DEBUG nova.network.neutron [req-75ee0a07-f207-436f-a46c-e1c0f30613e4 req-f3790d3f-29e2-4bbf-bf04-8acd1f9caf61 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1192.700615] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1192.700834] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1192.700993] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleting the datastore file [datastore2] 992920c3-af4f-489e-b21f-9d52c24399de {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1192.701577] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2af6760f-ea42-4124-a80a-57b2b9b55e08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.708902] env[65385]: DEBUG oslo_vmware.api [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for the task: (returnval){ [ 1192.708902] env[65385]: value = "task-4454640" [ 1192.708902] env[65385]: _type = "Task" [ 1192.708902] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.718971] env[65385]: DEBUG oslo_vmware.api [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.846521] env[65385]: DEBUG oslo_vmware.api [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454637, 'name': PowerOnVM_Task, 'duration_secs': 0.591413} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.847084] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1192.847626] env[65385]: INFO nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Took 10.24 seconds to spawn the instance on the hypervisor. [ 1192.847626] env[65385]: DEBUG nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1192.848824] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1d5fab-0837-4341-aca0-6e9c6d1a5d76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.108536] env[65385]: DEBUG nova.network.neutron [-] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1193.147539] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4fb7a59-349f-4e53-aa93-533bbbbd3ed4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.160478] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4801157-82a8-4159-ae2a-24a3b6ee2c85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.192762] env[65385]: DEBUG nova.compute.manager [req-75ee0a07-f207-436f-a46c-e1c0f30613e4 req-f3790d3f-29e2-4bbf-bf04-8acd1f9caf61 service nova] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Detach interface failed, port_id=b4e12be3-8c2c-49d7-8f67-781a1ee6680e, reason: Instance 9de8e3db-35b5-43a8-b677-d815055a0a51 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1193.220432] env[65385]: DEBUG oslo_vmware.api [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Task: {'id': task-4454640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.31956} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.220741] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1193.221016] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1193.221278] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1193.221505] env[65385]: INFO nova.compute.manager [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1193.221761] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1193.222041] env[65385]: DEBUG nova.compute.manager [-] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1193.222154] env[65385]: DEBUG nova.network.neutron [-] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1193.222456] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.223094] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.223484] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.313037] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.373407] env[65385]: INFO nova.compute.manager [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Took 18.91 seconds to build instance. [ 1193.611325] env[65385]: INFO nova.compute.manager [-] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Took 1.36 seconds to deallocate network for instance. [ 1193.638607] env[65385]: DEBUG nova.compute.manager [req-2c3f82eb-58bb-4c06-bb55-db59739f6202 req-dcf44bb0-0c8f-4d93-920f-79d6131ecc23 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Received event network-vif-deleted-62dbbb11-2d9c-4536-a2ed-e07b1e51dff7 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1193.639473] env[65385]: INFO nova.compute.manager [req-2c3f82eb-58bb-4c06-bb55-db59739f6202 req-dcf44bb0-0c8f-4d93-920f-79d6131ecc23 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Neutron deleted interface 62dbbb11-2d9c-4536-a2ed-e07b1e51dff7; detaching it from the instance and deleting it from the info cache [ 1193.639609] env[65385]: DEBUG nova.network.neutron [req-2c3f82eb-58bb-4c06-bb55-db59739f6202 req-dcf44bb0-0c8f-4d93-920f-79d6131ecc23 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1193.875362] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2572480e-eeb8-4cdc-b913-2bbb25c4f70a tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.420s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.034855] env[65385]: DEBUG nova.network.neutron [-] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.119966] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1194.120350] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1194.120658] env[65385]: DEBUG nova.objects.instance [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lazy-loading 'resources' on Instance uuid 9de8e3db-35b5-43a8-b677-d815055a0a51 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.142347] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3be8010d-7890-4c0c-86d5-e7ac9de720da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.154385] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910a3e68-54b6-4fff-b928-1ed9215eb138 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.190935] env[65385]: DEBUG nova.compute.manager [req-2c3f82eb-58bb-4c06-bb55-db59739f6202 req-dcf44bb0-0c8f-4d93-920f-79d6131ecc23 service nova] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Detach interface failed, port_id=62dbbb11-2d9c-4536-a2ed-e07b1e51dff7, reason: Instance 992920c3-af4f-489e-b21f-9d52c24399de could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1194.538717] env[65385]: INFO nova.compute.manager [-] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Took 1.32 seconds to deallocate network for instance. [ 1194.732402] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c9c324-ee6a-42f9-bc0f-9066070558ee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.741396] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2c121c-bca2-4224-9375-188d6e36c5ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.779193] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864943dd-6414-44b1-b6db-a62d40ca7318 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.788863] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76a1d33-691c-4f50-8bfd-25b0cd5c7271 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.804782] env[65385]: DEBUG nova.compute.provider_tree [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1195.051067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.308643] env[65385]: DEBUG nova.scheduler.client.report [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1195.664781] env[65385]: DEBUG nova.compute.manager [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Received event network-changed-34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1195.664942] env[65385]: DEBUG nova.compute.manager [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Refreshing instance network info cache due to event network-changed-34d33bf4-6698-4d35-a2ee-636a6e0fdbec. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1195.665171] env[65385]: DEBUG oslo_concurrency.lockutils [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Acquiring lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.665307] env[65385]: DEBUG oslo_concurrency.lockutils [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Acquired lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1195.665461] env[65385]: DEBUG nova.network.neutron [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Refreshing network info cache for port 34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1195.813567] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.816240] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.765s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.816475] env[65385]: DEBUG nova.objects.instance [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lazy-loading 'resources' on Instance uuid 992920c3-af4f-489e-b21f-9d52c24399de {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.839244] env[65385]: INFO nova.scheduler.client.report [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Deleted allocations for instance 9de8e3db-35b5-43a8-b677-d815055a0a51 [ 1196.168435] env[65385]: WARNING neutronclient.v2_0.client [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.169256] env[65385]: WARNING openstack [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.169607] env[65385]: WARNING openstack [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.294394] env[65385]: WARNING openstack [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.294880] env[65385]: WARNING openstack [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.347588] env[65385]: DEBUG oslo_concurrency.lockutils [None req-dcd385b3-000e-42fb-95b9-1caa251a04fa tempest-ServerDiskConfigTestJSON-1442739231 tempest-ServerDiskConfigTestJSON-1442739231-project-member] Lock "9de8e3db-35b5-43a8-b677-d815055a0a51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.756s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1196.356857] env[65385]: WARNING neutronclient.v2_0.client [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.357528] env[65385]: WARNING openstack [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.357934] env[65385]: WARNING openstack [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.415580] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d302fb07-35e4-4cf6-ac6f-9dc31ca7c69f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.424452] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd490b8-69b7-4ae0-975d-2f84b92bd3ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.459956] env[65385]: DEBUG nova.network.neutron [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updated VIF entry in instance network info cache for port 34d33bf4-6698-4d35-a2ee-636a6e0fdbec. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1196.460385] env[65385]: DEBUG nova.network.neutron [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updating instance_info_cache with network_info: [{"id": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "address": "fa:16:3e:83:87:18", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d33bf4-66", "ovs_interfaceid": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1196.462296] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5906744f-f6b0-461a-975e-7ae386fc4b85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.473569] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2ef674-f14c-4d58-a8ee-6f5a5b48bb85 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.490150] env[65385]: DEBUG nova.compute.provider_tree [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1196.965892] env[65385]: DEBUG oslo_concurrency.lockutils [req-2d94ae8f-7257-4d8d-a159-c17cf41a0aa8 req-bd6f2b26-edce-4a55-a370-747b6dc74720 service nova] Releasing lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.993359] env[65385]: DEBUG nova.scheduler.client.report [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1197.499100] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.527057] env[65385]: INFO nova.scheduler.client.report [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Deleted allocations for instance 992920c3-af4f-489e-b21f-9d52c24399de [ 1198.035676] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5676826-6a79-459f-aaeb-a852575c293f tempest-ServersTestJSON-776435614 tempest-ServersTestJSON-776435614-project-member] Lock "992920c3-af4f-489e-b21f-9d52c24399de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.473s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.635685] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "447d0bce-8995-4bef-bf1a-58269898a461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.636028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "447d0bce-8995-4bef-bf1a-58269898a461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1201.139338] env[65385]: DEBUG nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1201.665023] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1201.665363] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1201.666952] env[65385]: INFO nova.compute.claims [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1202.453776] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f7a834-9943-062d-32e0-7a84503af1ab/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1202.454782] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fd77a6-c13b-408f-a8b2-f861e995d265 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.461906] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f7a834-9943-062d-32e0-7a84503af1ab/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1202.462090] env[65385]: ERROR oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f7a834-9943-062d-32e0-7a84503af1ab/disk-0.vmdk due to incomplete transfer. [ 1202.462332] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e7f3a31b-b942-4bea-8559-2ffb41c2ffbe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.470343] env[65385]: DEBUG oslo_vmware.rw_handles [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f7a834-9943-062d-32e0-7a84503af1ab/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1202.470540] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Uploaded image d71dbee8-5170-4c28-88aa-44822c744198 to the Glance image server {{(pid=65385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1202.473305] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Destroying the VM {{(pid=65385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1202.473562] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-24b9f306-4938-45e9-bc73-9bd187d26d08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.480173] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1202.480173] env[65385]: value = "task-4454641" [ 1202.480173] env[65385]: _type = "Task" [ 1202.480173] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.488964] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454641, 'name': Destroy_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.766658] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945f7f6b-000f-4cb9-b7f2-ce5495923c45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.775340] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a5f637-1349-4756-808b-b90d42177515 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.807274] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65283823-8cec-451e-b1b7-ee99f7ecc57d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.815856] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c435fc69-8f24-497a-90b8-2eb121bb4b6c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.832271] env[65385]: DEBUG nova.compute.provider_tree [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.990890] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454641, 'name': Destroy_Task, 'duration_secs': 0.36361} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.991080] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Destroyed the VM [ 1202.991310] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Deleting Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1202.991603] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a2a85676-a8b6-49de-aa97-5dbb23c8fd79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.998629] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1202.998629] env[65385]: value = "task-4454642" [ 1202.998629] env[65385]: _type = "Task" [ 1202.998629] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.008976] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454642, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.336130] env[65385]: DEBUG nova.scheduler.client.report [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1203.509834] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454642, 'name': RemoveSnapshot_Task, 'duration_secs': 0.426863} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.510058] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Deleted Snapshot of the VM instance {{(pid=65385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1203.510365] env[65385]: DEBUG nova.compute.manager [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1203.511148] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2d81c1-518a-449f-8997-6936d0da351e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.841334] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.176s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.841828] env[65385]: DEBUG nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1204.023755] env[65385]: INFO nova.compute.manager [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Shelve offloading [ 1204.347351] env[65385]: DEBUG nova.compute.utils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1204.348915] env[65385]: DEBUG nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Not allocating networking since 'none' was specified. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 1204.527189] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1204.527516] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c72ceb4d-90d7-4b30-b303-6f759195cdae {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.535776] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1204.535776] env[65385]: value = "task-4454643" [ 1204.535776] env[65385]: _type = "Task" [ 1204.535776] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.543917] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454643, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.851418] env[65385]: DEBUG nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1205.048554] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1205.048755] env[65385]: DEBUG nova.compute.manager [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1205.049621] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556f75f3-c9e2-4b26-83b3-2a9c28bf867d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.056128] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.056292] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1205.056462] env[65385]: DEBUG nova.network.neutron [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1205.558872] env[65385]: WARNING neutronclient.v2_0.client [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.559621] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.559988] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.677090] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.677496] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.741976] env[65385]: WARNING neutronclient.v2_0.client [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.742671] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.743009] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.824911] env[65385]: DEBUG nova.network.neutron [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf3df048-71", "ovs_interfaceid": "cf3df048-713a-474d-9278-a467e50ec4f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1205.860223] env[65385]: DEBUG nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1205.887043] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1205.887298] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1205.887447] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1205.887621] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1205.887762] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1205.887904] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1205.888125] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1205.888279] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1205.888437] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1205.888594] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1205.888761] env[65385]: DEBUG nova.virt.hardware [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1205.889637] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b18952-96c2-4884-bb1a-b3df17391e09 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.898495] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf538472-839d-4759-96fa-97c7a4eddbac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.913787] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Instance VIF info [] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1205.919430] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Creating folder: Project (4e8b56786c834391b00d390afe1a9d8b). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1205.919706] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7293a3d4-dccc-48c3-a00e-d406c9bf0a70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.930518] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Created folder: Project (4e8b56786c834391b00d390afe1a9d8b) in parent group-v870881. [ 1205.930702] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Creating folder: Instances. Parent ref: group-v871201. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1205.930927] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1820dc40-45ac-4a73-a8b4-3c1c686ac64f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.940475] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Created folder: Instances in parent group-v871201. [ 1205.940695] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1205.940886] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1205.941106] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4058c511-97f5-4ebe-8f31-77eb1d3806b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.958765] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1205.958765] env[65385]: value = "task-4454646" [ 1205.958765] env[65385]: _type = "Task" [ 1205.958765] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.966852] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454646, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.327926] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1206.328405] env[65385]: WARNING neutronclient.v2_0.client [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.329074] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.329479] env[65385]: WARNING openstack [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.334580] env[65385]: WARNING neutronclient.v2_0.client [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.469238] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454646, 'name': CreateVM_Task, 'duration_secs': 0.39556} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.469441] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1206.469841] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.469996] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1206.470330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1206.470593] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82ae6bc8-a992-472a-8a19-9eeb1a9eb130 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.475615] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1206.475615] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520b89d8-ea6a-189e-03bb-7df391f56ad1" [ 1206.475615] env[65385]: _type = "Task" [ 1206.475615] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.483847] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520b89d8-ea6a-189e-03bb-7df391f56ad1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.563030] env[65385]: DEBUG nova.compute.manager [req-0e1cd090-151b-4654-9c68-2a852773acfe req-ec12545c-1d7d-471b-a864-a7c256cb6785 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-vif-unplugged-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1206.563417] env[65385]: DEBUG oslo_concurrency.lockutils [req-0e1cd090-151b-4654-9c68-2a852773acfe req-ec12545c-1d7d-471b-a864-a7c256cb6785 service nova] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.563417] env[65385]: DEBUG oslo_concurrency.lockutils [req-0e1cd090-151b-4654-9c68-2a852773acfe req-ec12545c-1d7d-471b-a864-a7c256cb6785 service nova] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.563758] env[65385]: DEBUG oslo_concurrency.lockutils [req-0e1cd090-151b-4654-9c68-2a852773acfe req-ec12545c-1d7d-471b-a864-a7c256cb6785 service nova] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.563820] env[65385]: DEBUG nova.compute.manager [req-0e1cd090-151b-4654-9c68-2a852773acfe req-ec12545c-1d7d-471b-a864-a7c256cb6785 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] No waiting events found dispatching network-vif-unplugged-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1206.563982] env[65385]: WARNING nova.compute.manager [req-0e1cd090-151b-4654-9c68-2a852773acfe req-ec12545c-1d7d-471b-a864-a7c256cb6785 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received unexpected event network-vif-unplugged-cf3df048-713a-474d-9278-a467e50ec4f8 for instance with vm_state shelved and task_state shelving_offloading. [ 1206.667410] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1206.668340] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7bfca0-d01d-4f6a-a10f-f57de1d48cdd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.677014] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1206.677338] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a9064ab-2021-41c1-a45b-e2ce9c7f021a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.750401] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1206.750616] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1206.750792] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleting the datastore file [datastore1] 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1206.751087] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5681b239-c802-4a3f-9a19-519cca76c2d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.759469] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1206.759469] env[65385]: value = "task-4454648" [ 1206.759469] env[65385]: _type = "Task" [ 1206.759469] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.769590] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454648, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.986666] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520b89d8-ea6a-189e-03bb-7df391f56ad1, 'name': SearchDatastore_Task, 'duration_secs': 0.009443} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.987056] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1206.987300] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1206.987584] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.987775] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1206.987964] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1206.988327] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da092348-19cd-4a85-b64e-fda78bfb4d63 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.997599] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1206.997749] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1206.998489] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7854468-e75d-4761-984e-bf044fa9ca34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.003657] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1207.003657] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522d32a2-b131-48d0-a6da-9bb21f85e3a4" [ 1207.003657] env[65385]: _type = "Task" [ 1207.003657] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.011444] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522d32a2-b131-48d0-a6da-9bb21f85e3a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.270203] env[65385]: DEBUG oslo_vmware.api [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454648, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160247} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.270495] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1207.270692] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1207.270863] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1207.295654] env[65385]: INFO nova.scheduler.client.report [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted allocations for instance 20933973-82ae-498c-b016-3a82bb5dc165 [ 1207.514952] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]522d32a2-b131-48d0-a6da-9bb21f85e3a4, 'name': SearchDatastore_Task, 'duration_secs': 0.00828} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.515790] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-864850d5-86e5-46e8-9974-c5eed101b06e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.521659] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1207.521659] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5209f3e4-047d-5242-f909-ba341b85aa5c" [ 1207.521659] env[65385]: _type = "Task" [ 1207.521659] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.529708] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5209f3e4-047d-5242-f909-ba341b85aa5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.800772] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1207.801097] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.801333] env[65385]: DEBUG nova.objects.instance [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'resources' on Instance uuid 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.033726] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5209f3e4-047d-5242-f909-ba341b85aa5c, 'name': SearchDatastore_Task, 'duration_secs': 0.010625} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.034176] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1208.034512] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 447d0bce-8995-4bef-bf1a-58269898a461/447d0bce-8995-4bef-bf1a-58269898a461.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1208.034862] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1bae6f9-7635-4f18-afac-4eecb2c92906 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.043653] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1208.043653] env[65385]: value = "task-4454649" [ 1208.043653] env[65385]: _type = "Task" [ 1208.043653] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.053431] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454649, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.304449] env[65385]: DEBUG nova.objects.instance [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'numa_topology' on Instance uuid 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.554366] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454649, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434345} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.554701] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 447d0bce-8995-4bef-bf1a-58269898a461/447d0bce-8995-4bef-bf1a-58269898a461.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1208.555146] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1208.555146] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-150ce614-814d-4bed-8e1a-93d239f3c9e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.562098] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1208.562098] env[65385]: value = "task-4454650" [ 1208.562098] env[65385]: _type = "Task" [ 1208.562098] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.571527] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454650, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.598064] env[65385]: DEBUG nova.compute.manager [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1208.598335] env[65385]: DEBUG nova.compute.manager [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing instance network info cache due to event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1208.598622] env[65385]: DEBUG oslo_concurrency.lockutils [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Acquiring lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.598768] env[65385]: DEBUG oslo_concurrency.lockutils [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Acquired lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1208.599009] env[65385]: DEBUG nova.network.neutron [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1208.807318] env[65385]: DEBUG nova.objects.base [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Object Instance<20933973-82ae-498c-b016-3a82bb5dc165> lazy-loaded attributes: resources,numa_topology {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1208.848357] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1208.891017] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27e4cb1-3e1f-4083-af3b-b071c9091f9f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.897877] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffa43ce-6060-439f-8f07-b5b1234825bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.932398] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccf66ec-615f-4f4a-bda1-3151439b37b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.941360] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6a340c-0d89-4f3d-bc6b-e2e73741dff7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.956673] env[65385]: DEBUG nova.compute.provider_tree [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.073050] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454650, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065155} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.073559] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1209.074217] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a5fd9c-2dc7-4130-8f11-20a5068b39b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.095377] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 447d0bce-8995-4bef-bf1a-58269898a461/447d0bce-8995-4bef-bf1a-58269898a461.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1209.096197] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-999b49ed-60f9-4dd0-95fa-64e356c24435 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.110803] env[65385]: WARNING neutronclient.v2_0.client [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.111892] env[65385]: WARNING openstack [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.112320] env[65385]: WARNING openstack [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.126022] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1209.126022] env[65385]: value = "task-4454651" [ 1209.126022] env[65385]: _type = "Task" [ 1209.126022] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.137010] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454651, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.238333] env[65385]: WARNING openstack [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.238757] env[65385]: WARNING openstack [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.303799] env[65385]: WARNING neutronclient.v2_0.client [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.304565] env[65385]: WARNING openstack [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.304932] env[65385]: WARNING openstack [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.384438] env[65385]: DEBUG nova.network.neutron [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updated VIF entry in instance network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1209.384831] env[65385]: DEBUG nova.network.neutron [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": null, "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapcf3df048-71", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1209.459966] env[65385]: DEBUG nova.scheduler.client.report [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1209.636938] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454651, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.887563] env[65385]: DEBUG oslo_concurrency.lockutils [req-33acfce2-fe2d-45eb-b4ce-b2007725ffa5 req-40aa8227-c95b-47e0-837a-c8a6bde8dfa1 service nova] Releasing lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1209.965774] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.165s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.137790] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454651, 'name': ReconfigVM_Task, 'duration_secs': 0.708546} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.138105] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 447d0bce-8995-4bef-bf1a-58269898a461/447d0bce-8995-4bef-bf1a-58269898a461.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1210.138680] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-833dea90-4813-4295-94c7-9e225c17ea3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.146464] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1210.146464] env[65385]: value = "task-4454652" [ 1210.146464] env[65385]: _type = "Task" [ 1210.146464] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.154744] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454652, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.474641] env[65385]: DEBUG oslo_concurrency.lockutils [None req-798f1fa2-f7ba-4231-a6e9-57794c250a56 tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.034s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.476028] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.628s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.476185] env[65385]: INFO nova.compute.manager [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Unshelving [ 1210.657816] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454652, 'name': Rename_Task, 'duration_secs': 0.129222} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.658115] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1210.658385] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f533632-322f-4ce2-ac46-cd6a1da1bc52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.667080] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1210.667080] env[65385]: value = "task-4454653" [ 1210.667080] env[65385]: _type = "Task" [ 1210.667080] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.674853] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454653, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.177586] env[65385]: DEBUG oslo_vmware.api [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454653, 'name': PowerOnVM_Task, 'duration_secs': 0.457517} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.178032] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1211.178280] env[65385]: INFO nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Took 5.32 seconds to spawn the instance on the hypervisor. [ 1211.178537] env[65385]: DEBUG nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1211.179402] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5003b79a-97c1-488a-a2a5-9f5cb41416b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.504904] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.505232] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.505438] env[65385]: DEBUG nova.objects.instance [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'pci_requests' on Instance uuid 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.698047] env[65385]: INFO nova.compute.manager [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Took 10.05 seconds to build instance. [ 1212.008922] env[65385]: DEBUG nova.objects.instance [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'numa_topology' on Instance uuid 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1212.200957] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6ab472b-4dc7-4d1d-959f-ab04c0fe27f6 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "447d0bce-8995-4bef-bf1a-58269898a461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.565s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1212.511409] env[65385]: INFO nova.compute.claims [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1212.672864] env[65385]: DEBUG nova.compute.manager [None req-5a8d8c57-324d-4388-94d1-ae925bc09e05 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1212.673883] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e2ba83-18bf-4310-8742-32a6d48e7ead {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.755922] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "447d0bce-8995-4bef-bf1a-58269898a461" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1212.756324] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "447d0bce-8995-4bef-bf1a-58269898a461" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1212.756621] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "447d0bce-8995-4bef-bf1a-58269898a461-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1212.756912] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "447d0bce-8995-4bef-bf1a-58269898a461-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1212.757178] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "447d0bce-8995-4bef-bf1a-58269898a461-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1212.760065] env[65385]: INFO nova.compute.manager [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Terminating instance [ 1213.184722] env[65385]: INFO nova.compute.manager [None req-5a8d8c57-324d-4388-94d1-ae925bc09e05 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] instance snapshotting [ 1213.185392] env[65385]: DEBUG nova.objects.instance [None req-5a8d8c57-324d-4388-94d1-ae925bc09e05 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lazy-loading 'flavor' on Instance uuid 447d0bce-8995-4bef-bf1a-58269898a461 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.265150] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "refresh_cache-447d0bce-8995-4bef-bf1a-58269898a461" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.265360] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquired lock "refresh_cache-447d0bce-8995-4bef-bf1a-58269898a461" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.265402] env[65385]: DEBUG nova.network.neutron [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1213.613581] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d5680e-9090-493b-a43a-08d20f47189c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.621995] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d1b2e2-d23b-43d8-b07a-52c74801d491 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.652893] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0406bd-a982-4aa1-89b5-8bb1d5be5608 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.661190] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f1fcc6-91c0-44ee-9ef3-d9e191f9d93a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.675022] env[65385]: DEBUG nova.compute.provider_tree [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.694935] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18daa257-092f-4b8e-a0a6-be029b562f75 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.721293] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccec63b-e523-4d89-81d5-849491b14025 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.768442] env[65385]: WARNING neutronclient.v2_0.client [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.769315] env[65385]: WARNING openstack [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.769793] env[65385]: WARNING openstack [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.796084] env[65385]: DEBUG nova.network.neutron [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1213.859312] env[65385]: DEBUG nova.network.neutron [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1214.177992] env[65385]: DEBUG nova.scheduler.client.report [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1214.236075] env[65385]: DEBUG nova.compute.manager [None req-5a8d8c57-324d-4388-94d1-ae925bc09e05 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Instance disappeared during snapshot {{(pid=65385) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4635}} [ 1214.357594] env[65385]: DEBUG nova.compute.manager [None req-5a8d8c57-324d-4388-94d1-ae925bc09e05 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Found 0 images (rotation: 2) {{(pid=65385) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 1214.361459] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Releasing lock "refresh_cache-447d0bce-8995-4bef-bf1a-58269898a461" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1214.361900] env[65385]: DEBUG nova.compute.manager [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1214.362110] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1214.362999] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef3ce0e-aee1-4ebf-9b6a-85d9badf45a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.370950] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1214.371241] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44a20706-546c-40da-a4fe-fa52cd6b30fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.378300] env[65385]: DEBUG oslo_vmware.api [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1214.378300] env[65385]: value = "task-4454654" [ 1214.378300] env[65385]: _type = "Task" [ 1214.378300] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.387584] env[65385]: DEBUG oslo_vmware.api [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.685071] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.177s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1214.685071] env[65385]: WARNING neutronclient.v2_0.client [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.725121] env[65385]: INFO nova.network.neutron [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating port cf3df048-713a-474d-9278-a467e50ec4f8 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1214.889044] env[65385]: DEBUG oslo_vmware.api [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454654, 'name': PowerOffVM_Task, 'duration_secs': 0.176731} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.889397] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1214.889562] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1214.889819] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9afad6e-0844-4083-9b34-0dc6604212c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.919521] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1214.919792] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1214.919990] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Deleting the datastore file [datastore2] 447d0bce-8995-4bef-bf1a-58269898a461 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1214.920286] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0292b14-ba0f-42df-abf6-1293cc8413c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.927713] env[65385]: DEBUG oslo_vmware.api [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for the task: (returnval){ [ 1214.927713] env[65385]: value = "task-4454656" [ 1214.927713] env[65385]: _type = "Task" [ 1214.927713] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.937216] env[65385]: DEBUG oslo_vmware.api [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.438616] env[65385]: DEBUG oslo_vmware.api [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Task: {'id': task-4454656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10368} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.439089] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1215.439089] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1215.439231] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1215.439391] env[65385]: INFO nova.compute.manager [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1215.439628] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1215.439822] env[65385]: DEBUG nova.compute.manager [-] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1215.439919] env[65385]: DEBUG nova.network.neutron [-] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1215.440174] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.440695] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1215.440951] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1215.462175] env[65385]: DEBUG nova.network.neutron [-] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1215.462432] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.965128] env[65385]: DEBUG nova.network.neutron [-] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1216.174959] env[65385]: DEBUG nova.compute.manager [req-561879fd-4b9e-4d37-b9d1-3d5f7b6ae176 req-f5d0cadc-24f4-4faa-96ab-015d9ba2a339 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-vif-plugged-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1216.175125] env[65385]: DEBUG oslo_concurrency.lockutils [req-561879fd-4b9e-4d37-b9d1-3d5f7b6ae176 req-f5d0cadc-24f4-4faa-96ab-015d9ba2a339 service nova] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1216.175396] env[65385]: DEBUG oslo_concurrency.lockutils [req-561879fd-4b9e-4d37-b9d1-3d5f7b6ae176 req-f5d0cadc-24f4-4faa-96ab-015d9ba2a339 service nova] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1216.175674] env[65385]: DEBUG oslo_concurrency.lockutils [req-561879fd-4b9e-4d37-b9d1-3d5f7b6ae176 req-f5d0cadc-24f4-4faa-96ab-015d9ba2a339 service nova] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1216.175888] env[65385]: DEBUG nova.compute.manager [req-561879fd-4b9e-4d37-b9d1-3d5f7b6ae176 req-f5d0cadc-24f4-4faa-96ab-015d9ba2a339 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] No waiting events found dispatching network-vif-plugged-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1216.176108] env[65385]: WARNING nova.compute.manager [req-561879fd-4b9e-4d37-b9d1-3d5f7b6ae176 req-f5d0cadc-24f4-4faa-96ab-015d9ba2a339 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received unexpected event network-vif-plugged-cf3df048-713a-474d-9278-a467e50ec4f8 for instance with vm_state shelved_offloaded and task_state spawning. [ 1216.238202] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.238430] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1216.238572] env[65385]: DEBUG nova.network.neutron [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1216.467769] env[65385]: INFO nova.compute.manager [-] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Took 1.03 seconds to deallocate network for instance. [ 1216.742770] env[65385]: WARNING neutronclient.v2_0.client [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1216.742770] env[65385]: WARNING openstack [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.743182] env[65385]: WARNING openstack [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.859538] env[65385]: WARNING openstack [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.859967] env[65385]: WARNING openstack [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.923686] env[65385]: WARNING neutronclient.v2_0.client [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1216.924447] env[65385]: WARNING openstack [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.924894] env[65385]: WARNING openstack [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.973686] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1216.974053] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1216.974332] env[65385]: DEBUG nova.objects.instance [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lazy-loading 'resources' on Instance uuid 447d0bce-8995-4bef-bf1a-58269898a461 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.006849] env[65385]: DEBUG nova.network.neutron [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf3df048-71", "ovs_interfaceid": "cf3df048-713a-474d-9278-a467e50ec4f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1217.508995] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1217.537402] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1d04ba74ba362e299d0574f6f8ce371d',container_format='bare',created_at=2025-11-14T16:57:01Z,direct_url=,disk_format='vmdk',id=d71dbee8-5170-4c28-88aa-44822c744198,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-674811985-shelved',owner='1598307c304d46a884a88b3eb5f5d7dc',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2025-11-14T16:57:22Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1217.537645] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1217.537795] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1217.537972] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1217.538129] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1217.538268] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1217.538473] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1217.538627] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1217.538785] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1217.538938] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1217.539112] env[65385]: DEBUG nova.virt.hardware [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1217.539963] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567fd5ec-177e-4bd1-aca6-9a6684769811 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.551016] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d6b5c0-d8bd-46bb-b1b1-9de8f56676ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.567238] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:47:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9875d38f-76e2-416c-bfb7-f18a22b0d8ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf3df048-713a-474d-9278-a467e50ec4f8', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1217.574619] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1217.577134] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1217.577524] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35471c5e-12fe-45ac-b5ee-356b8d716ee7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.595343] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3510f5-27b6-4443-8f7c-6d1e4ac8c931 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.604922] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e892120-ec40-4c6e-8a66-adbab9efe474 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.607979] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1217.607979] env[65385]: value = "task-4454657" [ 1217.607979] env[65385]: _type = "Task" [ 1217.607979] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.636754] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c8ce45-d95a-4c90-b904-45768285d374 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.642764] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454657, 'name': CreateVM_Task} progress is 15%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.648219] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7986baee-5238-4326-9a15-9f469b5ebcc4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.664331] env[65385]: DEBUG nova.compute.provider_tree [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.119221] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454657, 'name': CreateVM_Task, 'duration_secs': 0.314266} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.119384] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1218.119852] env[65385]: WARNING neutronclient.v2_0.client [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.120232] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.120380] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.120748] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1218.121022] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a5eb156-a6ba-4627-8ce8-ab0c32276c70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.127192] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1218.127192] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521d85f3-9700-939a-6b54-f4a43b14da7b" [ 1218.127192] env[65385]: _type = "Task" [ 1218.127192] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.137019] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521d85f3-9700-939a-6b54-f4a43b14da7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.168401] env[65385]: DEBUG nova.scheduler.client.report [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1218.201012] env[65385]: DEBUG nova.compute.manager [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1218.201222] env[65385]: DEBUG nova.compute.manager [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing instance network info cache due to event network-changed-cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1218.201434] env[65385]: DEBUG oslo_concurrency.lockutils [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Acquiring lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.201570] env[65385]: DEBUG oslo_concurrency.lockutils [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Acquired lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.201730] env[65385]: DEBUG nova.network.neutron [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Refreshing network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1218.638919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1218.639354] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Processing image d71dbee8-5170-4c28-88aa-44822c744198 {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1218.639405] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198/d71dbee8-5170-4c28-88aa-44822c744198.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.639524] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198/d71dbee8-5170-4c28-88aa-44822c744198.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.639697] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1218.639953] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4637c712-fd40-48b9-8ae6-9ef729329256 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.659295] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1218.659480] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1218.660189] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c478aa0-0d69-4c8a-abf5-54e1f8dc4a19 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.665720] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1218.665720] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5288b8a8-2327-d0af-2b88-ecf7820432e9" [ 1218.665720] env[65385]: _type = "Task" [ 1218.665720] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.674151] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.676037] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5288b8a8-2327-d0af-2b88-ecf7820432e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.693358] env[65385]: INFO nova.scheduler.client.report [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Deleted allocations for instance 447d0bce-8995-4bef-bf1a-58269898a461 [ 1218.704233] env[65385]: WARNING neutronclient.v2_0.client [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.704914] env[65385]: WARNING openstack [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.705293] env[65385]: WARNING openstack [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.835178] env[65385]: WARNING openstack [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.835590] env[65385]: WARNING openstack [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.896346] env[65385]: WARNING neutronclient.v2_0.client [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.896986] env[65385]: WARNING openstack [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.897346] env[65385]: WARNING openstack [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.976438] env[65385]: DEBUG nova.network.neutron [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updated VIF entry in instance network info cache for port cf3df048-713a-474d-9278-a467e50ec4f8. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1218.976819] env[65385]: DEBUG nova.network.neutron [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [{"id": "cf3df048-713a-474d-9278-a467e50ec4f8", "address": "fa:16:3e:f4:47:b8", "network": {"id": "89c4ad71-8b83-4750-a4df-8cef16029037", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2130485540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1598307c304d46a884a88b3eb5f5d7dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf3df048-71", "ovs_interfaceid": "cf3df048-713a-474d-9278-a467e50ec4f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1219.177549] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Preparing fetch location {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1219.177831] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Fetch image to [datastore2] OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a/OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a.vmdk {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1219.178026] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Downloading stream optimized image d71dbee8-5170-4c28-88aa-44822c744198 to [datastore2] OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a/OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a.vmdk on the data store datastore2 as vApp {{(pid=65385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1219.178203] env[65385]: DEBUG nova.virt.vmwareapi.images [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Downloading image file data d71dbee8-5170-4c28-88aa-44822c744198 to the ESX as VM named 'OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a' {{(pid=65385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1219.202009] env[65385]: DEBUG oslo_concurrency.lockutils [None req-89b0e80e-9f57-4356-8959-93e53ae97b42 tempest-ServersAaction247Test-411426278 tempest-ServersAaction247Test-411426278-project-member] Lock "447d0bce-8995-4bef-bf1a-58269898a461" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.446s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.257729] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1219.257729] env[65385]: value = "resgroup-9" [ 1219.257729] env[65385]: _type = "ResourcePool" [ 1219.257729] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1219.258093] env[65385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8c4c186b-4cdc-4849-bbba-84a067d9c57d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.280208] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease: (returnval){ [ 1219.280208] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52391cc9-6bb6-7697-fe07-96a47b4dfe19" [ 1219.280208] env[65385]: _type = "HttpNfcLease" [ 1219.280208] env[65385]: } obtained for vApp import into resource pool (val){ [ 1219.280208] env[65385]: value = "resgroup-9" [ 1219.280208] env[65385]: _type = "ResourcePool" [ 1219.280208] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1219.280473] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the lease: (returnval){ [ 1219.280473] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52391cc9-6bb6-7697-fe07-96a47b4dfe19" [ 1219.280473] env[65385]: _type = "HttpNfcLease" [ 1219.280473] env[65385]: } to be ready. {{(pid=65385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1219.289967] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1219.289967] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52391cc9-6bb6-7697-fe07-96a47b4dfe19" [ 1219.289967] env[65385]: _type = "HttpNfcLease" [ 1219.289967] env[65385]: } is initializing. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1219.479987] env[65385]: DEBUG oslo_concurrency.lockutils [req-7661479e-3907-4d60-b0ee-dbd30c46c32e req-c84da8f7-3d0b-44c4-9c71-14929e0017db service nova] Releasing lock "refresh_cache-20933973-82ae-498c-b016-3a82bb5dc165" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1219.789467] env[65385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1219.789467] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52391cc9-6bb6-7697-fe07-96a47b4dfe19" [ 1219.789467] env[65385]: _type = "HttpNfcLease" [ 1219.789467] env[65385]: } is ready. {{(pid=65385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1219.789919] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1219.789919] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52391cc9-6bb6-7697-fe07-96a47b4dfe19" [ 1219.789919] env[65385]: _type = "HttpNfcLease" [ 1219.789919] env[65385]: }. {{(pid=65385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1219.790502] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11f9736-185c-402d-bf69-cf15da68ff73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.799257] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5277080a-7daf-e836-ebe1-7db3caf10310/disk-0.vmdk from lease info. {{(pid=65385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1219.799453] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5277080a-7daf-e836-ebe1-7db3caf10310/disk-0.vmdk. {{(pid=65385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1219.862971] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d76b5c3b-8a23-4aee-bc06-1a138f4077f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.932784] env[65385]: DEBUG oslo_concurrency.lockutils [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.933164] env[65385]: DEBUG oslo_concurrency.lockutils [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.989131] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Completed reading data from the image iterator. {{(pid=65385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1220.989667] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5277080a-7daf-e836-ebe1-7db3caf10310/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1220.990448] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac298ae2-8dca-45fa-8e7b-6843d5d350e7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.998227] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5277080a-7daf-e836-ebe1-7db3caf10310/disk-0.vmdk is in state: ready. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1220.998507] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5277080a-7daf-e836-ebe1-7db3caf10310/disk-0.vmdk. {{(pid=65385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1220.998709] env[65385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-7d4b1435-dde4-4495-8ab4-fa44ccd8b4f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.172057] env[65385]: DEBUG oslo_vmware.rw_handles [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5277080a-7daf-e836-ebe1-7db3caf10310/disk-0.vmdk. {{(pid=65385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1221.172057] env[65385]: INFO nova.virt.vmwareapi.images [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Downloaded image file data d71dbee8-5170-4c28-88aa-44822c744198 [ 1221.172914] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a562b3-203c-43f2-8a6c-40e725d880e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.189469] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-665629ca-6c88-4aa9-a689-b6495bab908a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.220711] env[65385]: INFO nova.virt.vmwareapi.images [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] The imported VM was unregistered [ 1221.223629] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Caching image {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1221.224073] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198 {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1221.224403] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bf6030a-7b65-44f0-b6e3-76990edd9101 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.236190] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Created directory with path [datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198 {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1221.236265] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a/OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a.vmdk to [datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198/d71dbee8-5170-4c28-88aa-44822c744198.vmdk. {{(pid=65385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1221.236542] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7b8f6e70-b804-4cbf-bb2e-5d4498e736ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.243768] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1221.243768] env[65385]: value = "task-4454660" [ 1221.243768] env[65385]: _type = "Task" [ 1221.243768] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.255996] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.436749] env[65385]: INFO nova.compute.manager [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Detaching volume c9bf37ba-bb9d-492e-98bf-bd450ee85fbc [ 1221.477255] env[65385]: INFO nova.virt.block_device [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Attempting to driver detach volume c9bf37ba-bb9d-492e-98bf-bd450ee85fbc from mountpoint /dev/sdb [ 1221.477524] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1221.477741] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1221.478785] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b378f64-ab6a-44d9-b1e0-b55662bacf78 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.503425] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3efd2b-821a-4ccf-a6a7-fcae2b2299b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.512207] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc475fc-3b6c-4144-a20e-e63f653995aa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.538687] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af5274a-714c-4d8e-8dce-1887ec233804 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.556345] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The volume has not been displaced from its original location: [datastore1] volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc/volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1221.562152] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1221.562598] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a086d20-0aae-4e34-ad1d-ed4ff600fe0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.582967] env[65385]: DEBUG oslo_vmware.api [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1221.582967] env[65385]: value = "task-4454661" [ 1221.582967] env[65385]: _type = "Task" [ 1221.582967] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.593373] env[65385]: DEBUG oslo_vmware.api [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.755262] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.094751] env[65385]: DEBUG oslo_vmware.api [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454661, 'name': ReconfigVM_Task, 'duration_secs': 0.318348} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.095248] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1222.100320] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06ae6ee7-97bf-4372-a855-08d74003383d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.118118] env[65385]: DEBUG oslo_vmware.api [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1222.118118] env[65385]: value = "task-4454662" [ 1222.118118] env[65385]: _type = "Task" [ 1222.118118] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.128146] env[65385]: DEBUG oslo_vmware.api [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454662, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.255740] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.628572] env[65385]: DEBUG oslo_vmware.api [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454662, 'name': ReconfigVM_Task, 'duration_secs': 0.167585} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.628829] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871192', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'name': 'volume-c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '62c30ded-c851-477f-b1d4-921268a6ef1c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc', 'serial': 'c9bf37ba-bb9d-492e-98bf-bd450ee85fbc'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1222.756659] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.184010] env[65385]: DEBUG nova.objects.instance [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'flavor' on Instance uuid 62c30ded-c851-477f-b1d4-921268a6ef1c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1223.258128] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.758176] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.192066] env[65385]: DEBUG oslo_concurrency.lockutils [None req-17632fea-9493-4a4e-8ee8-239f8dc03af6 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.258s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.258341] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.759587] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454660, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.411885} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.760086] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a/OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a.vmdk to [datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198/d71dbee8-5170-4c28-88aa-44822c744198.vmdk. [ 1224.760289] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Cleaning up location [datastore2] OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1224.760455] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_2c48f873-42e1-4bcf-a015-15fa2a46560a {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1224.760713] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fa77b07-8115-4672-b5ef-40ddd617fbf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.767556] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1224.767556] env[65385]: value = "task-4454663" [ 1224.767556] env[65385]: _type = "Task" [ 1224.767556] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.776465] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.194826] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.195212] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.195304] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "62c30ded-c851-477f-b1d4-921268a6ef1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.195482] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.195648] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.198011] env[65385]: INFO nova.compute.manager [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Terminating instance [ 1225.278427] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03499} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.278655] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1225.278815] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198/d71dbee8-5170-4c28-88aa-44822c744198.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.279077] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198/d71dbee8-5170-4c28-88aa-44822c744198.vmdk to [datastore2] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1225.279341] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-080c0689-b35f-4490-8917-61589b819691 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.286591] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1225.286591] env[65385]: value = "task-4454664" [ 1225.286591] env[65385]: _type = "Task" [ 1225.286591] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.295197] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.702342] env[65385]: DEBUG nova.compute.manager [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1225.702572] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1225.703510] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1088f8db-6efa-4d57-9416-ca3001cd2f1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.712177] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1225.712477] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5ab4ae2-de96-4957-977a-7be5b508bbf1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.719658] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1225.719658] env[65385]: value = "task-4454665" [ 1225.719658] env[65385]: _type = "Task" [ 1225.719658] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.730533] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.797968] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454664, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.231448] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454665, 'name': PowerOffVM_Task, 'duration_secs': 0.219966} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.231916] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1226.231962] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1226.232229] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfa4f97e-72d1-464a-9cd7-b8ea4f283a67 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.307294] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454664, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.312254] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1226.312475] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1226.312664] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleting the datastore file [datastore2] 62c30ded-c851-477f-b1d4-921268a6ef1c {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1226.312965] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-881edb43-0a65-4541-a7ab-586d0bd499d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.321752] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1226.321752] env[65385]: value = "task-4454667" [ 1226.321752] env[65385]: _type = "Task" [ 1226.321752] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.331463] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.797819] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454664, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.835733] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.299684] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454664, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.333285] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.801062] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454664, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.320522} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.801249] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d71dbee8-5170-4c28-88aa-44822c744198/d71dbee8-5170-4c28-88aa-44822c744198.vmdk to [datastore2] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1227.801944] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0659612f-ba4b-4e77-bb2b-03fa9b490583 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.825112] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1227.826796] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28bbaa94-ef49-40df-b0fa-93c459564213 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.854463] env[65385]: DEBUG oslo_vmware.api [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.316219} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.856712] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1227.856917] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1227.857111] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1227.857284] env[65385]: INFO nova.compute.manager [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Took 2.15 seconds to destroy the instance on the hypervisor. [ 1227.857618] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1227.858175] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1227.858175] env[65385]: value = "task-4454668" [ 1227.858175] env[65385]: _type = "Task" [ 1227.858175] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.858384] env[65385]: DEBUG nova.compute.manager [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1227.858487] env[65385]: DEBUG nova.network.neutron [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1227.858725] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1227.859266] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1227.859523] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1227.876619] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454668, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.920313] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1228.348029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1228.348029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.348029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1228.348029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.348029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1228.349629] env[65385]: INFO nova.compute.manager [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Terminating instance [ 1228.377971] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454668, 'name': ReconfigVM_Task, 'duration_secs': 0.388942} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.379676] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 20933973-82ae-498c-b016-3a82bb5dc165/20933973-82ae-498c-b016-3a82bb5dc165.vmdk or device None with type streamOptimized {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1228.381189] env[65385]: DEBUG nova.compute.manager [req-ba9156d0-8f20-4123-8092-8175138eef5d req-de34849c-e836-446f-8c11-fa5a0a9b3ec6 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Received event network-vif-deleted-e5e55898-473b-4cdf-9d34-0b704e923915 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1228.381379] env[65385]: INFO nova.compute.manager [req-ba9156d0-8f20-4123-8092-8175138eef5d req-de34849c-e836-446f-8c11-fa5a0a9b3ec6 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Neutron deleted interface e5e55898-473b-4cdf-9d34-0b704e923915; detaching it from the instance and deleting it from the info cache [ 1228.381549] env[65385]: DEBUG nova.network.neutron [req-ba9156d0-8f20-4123-8092-8175138eef5d req-de34849c-e836-446f-8c11-fa5a0a9b3ec6 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1228.383051] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf9fb8c4-2624-4660-be14-ceb47f55c8f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.391273] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1228.391273] env[65385]: value = "task-4454669" [ 1228.391273] env[65385]: _type = "Task" [ 1228.391273] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.401830] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454669, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.855191] env[65385]: DEBUG nova.compute.manager [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1228.855437] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1228.856358] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc377f0-f5a0-4ee5-a698-78a9e11ef52a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.859805] env[65385]: DEBUG nova.network.neutron [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1228.866502] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1228.866764] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51ee8a77-41a0-432e-8e95-5ae04bba76c6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.873820] env[65385]: DEBUG oslo_vmware.api [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1228.873820] env[65385]: value = "task-4454670" [ 1228.873820] env[65385]: _type = "Task" [ 1228.873820] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.883896] env[65385]: DEBUG oslo_vmware.api [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.884166] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-57f183a8-9acc-4d38-8a73-ba21fedd2d5d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.893494] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9861410a-9628-4612-ab1f-1cd0639b75fd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.916835] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454669, 'name': Rename_Task, 'duration_secs': 0.156663} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.917134] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1228.917388] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c0b65f6-9ccd-4d83-a1ce-2d61fc8d02d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.930637] env[65385]: DEBUG nova.compute.manager [req-ba9156d0-8f20-4123-8092-8175138eef5d req-de34849c-e836-446f-8c11-fa5a0a9b3ec6 service nova] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Detach interface failed, port_id=e5e55898-473b-4cdf-9d34-0b704e923915, reason: Instance 62c30ded-c851-477f-b1d4-921268a6ef1c could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1228.932363] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1228.932363] env[65385]: value = "task-4454671" [ 1228.932363] env[65385]: _type = "Task" [ 1228.932363] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.942512] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454671, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.363116] env[65385]: INFO nova.compute.manager [-] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Took 1.50 seconds to deallocate network for instance. [ 1229.386113] env[65385]: DEBUG oslo_vmware.api [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454670, 'name': PowerOffVM_Task, 'duration_secs': 0.20952} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.386398] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1229.386574] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1229.386843] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3a3b22d-2d8e-4b71-9577-487ca1c6d4be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.443054] env[65385]: DEBUG oslo_vmware.api [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454671, 'name': PowerOnVM_Task, 'duration_secs': 0.450179} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.443332] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1229.455975] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1229.456205] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1229.456332] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleting the datastore file [datastore1] df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1229.456614] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b39d379-e851-4567-aa63-43ee71050927 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.463920] env[65385]: DEBUG oslo_vmware.api [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for the task: (returnval){ [ 1229.463920] env[65385]: value = "task-4454673" [ 1229.463920] env[65385]: _type = "Task" [ 1229.463920] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.473635] env[65385]: DEBUG oslo_vmware.api [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.546258] env[65385]: DEBUG nova.compute.manager [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1229.547409] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a537bd8a-9da9-4df4-956b-7a183e140b79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.870818] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1229.871127] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1229.871359] env[65385]: DEBUG nova.objects.instance [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'resources' on Instance uuid 62c30ded-c851-477f-b1d4-921268a6ef1c {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.975978] env[65385]: DEBUG oslo_vmware.api [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Task: {'id': task-4454673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139202} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.976349] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1229.976603] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1229.976903] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1229.977128] env[65385]: INFO nova.compute.manager [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1229.977435] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1229.977692] env[65385]: DEBUG nova.compute.manager [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1229.977841] env[65385]: DEBUG nova.network.neutron [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1229.978159] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1229.978759] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1229.979135] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1230.033955] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1230.067602] env[65385]: DEBUG oslo_concurrency.lockutils [None req-431f43c3-6171-49c5-9bc6-177244553bef tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.592s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.467821] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e605db3-ee38-45bf-a491-e7353707c850 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.476900] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfa9b3a-1c3e-448b-9e5f-307fcf00f80c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.512848] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-211458f2-3ea8-4214-837a-0bf5f351f661 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.518942] env[65385]: DEBUG nova.compute.manager [req-8fd94af7-b394-421f-98d2-2a1ff2d92da8 req-5175218b-aef2-4bad-9c97-9715862f621d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Received event network-vif-deleted-cd731def-a6f7-40f9-b6ad-a10d2fd72984 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1230.519150] env[65385]: INFO nova.compute.manager [req-8fd94af7-b394-421f-98d2-2a1ff2d92da8 req-5175218b-aef2-4bad-9c97-9715862f621d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Neutron deleted interface cd731def-a6f7-40f9-b6ad-a10d2fd72984; detaching it from the instance and deleting it from the info cache [ 1230.519312] env[65385]: DEBUG nova.network.neutron [req-8fd94af7-b394-421f-98d2-2a1ff2d92da8 req-5175218b-aef2-4bad-9c97-9715862f621d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1230.527548] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b808c0-2e0b-416c-87ef-65fb6a75dc10 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.544705] env[65385]: DEBUG nova.compute.provider_tree [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1230.978052] env[65385]: DEBUG nova.network.neutron [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1231.020889] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1231.021212] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1231.021443] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "20933973-82ae-498c-b016-3a82bb5dc165-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1231.021630] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1231.021797] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1231.023885] env[65385]: INFO nova.compute.manager [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Terminating instance [ 1231.025151] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07e715bb-d2e3-40b4-892c-34ecf7a464cc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.036583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e28dea7-a274-4fb1-a381-44413d67c3eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.049352] env[65385]: DEBUG nova.scheduler.client.report [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1231.071730] env[65385]: DEBUG nova.compute.manager [req-8fd94af7-b394-421f-98d2-2a1ff2d92da8 req-5175218b-aef2-4bad-9c97-9715862f621d service nova] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Detach interface failed, port_id=cd731def-a6f7-40f9-b6ad-a10d2fd72984, reason: Instance df419705-e7e8-47b4-b9a0-2f1cf2638f33 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1231.480415] env[65385]: INFO nova.compute.manager [-] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Took 1.50 seconds to deallocate network for instance. [ 1231.529707] env[65385]: DEBUG nova.compute.manager [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1231.529875] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1231.530727] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a5e537-6a75-49cb-8888-7049dbecbfc4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.538914] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1231.539168] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e6c4c7e-e453-4442-b310-831bf51bd740 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.545803] env[65385]: DEBUG oslo_vmware.api [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1231.545803] env[65385]: value = "task-4454674" [ 1231.545803] env[65385]: _type = "Task" [ 1231.545803] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.555304] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1231.557349] env[65385]: DEBUG oslo_vmware.api [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.582123] env[65385]: INFO nova.scheduler.client.report [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted allocations for instance 62c30ded-c851-477f-b1d4-921268a6ef1c [ 1231.987686] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1231.988039] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1231.988257] env[65385]: DEBUG nova.objects.instance [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lazy-loading 'resources' on Instance uuid df419705-e7e8-47b4-b9a0-2f1cf2638f33 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.057067] env[65385]: DEBUG oslo_vmware.api [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454674, 'name': PowerOffVM_Task, 'duration_secs': 0.208213} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.057289] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1232.057464] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1232.057771] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfd348b4-132e-44c1-85c8-06e190e7b32b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.091251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-933f8603-2d14-40dd-80f0-1576be283d84 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "62c30ded-c851-477f-b1d4-921268a6ef1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.896s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1232.106660] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1232.106922] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1232.107137] env[65385]: DEBUG nova.compute.manager [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1232.108413] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89131c2d-4d67-43e6-9f75-5e691ca35099 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.116394] env[65385]: DEBUG nova.compute.manager [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1232.116964] env[65385]: DEBUG nova.objects.instance [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'flavor' on Instance uuid fa02564b-c1a8-4f8f-876c-806618e37f5b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.125172] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1232.125400] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1232.125619] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleting the datastore file [datastore2] 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1232.125912] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2cafc221-364f-4570-8af3-8a2367bea85e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.133172] env[65385]: DEBUG oslo_vmware.api [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for the task: (returnval){ [ 1232.133172] env[65385]: value = "task-4454676" [ 1232.133172] env[65385]: _type = "Task" [ 1232.133172] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.142705] env[65385]: DEBUG oslo_vmware.api [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.560742] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bc34c3-baeb-4b2d-b1ee-80bb7153ae5b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.568922] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fd2fe0-094c-461f-9523-5bba58b3704d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.601725] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af799481-eb55-4035-be2f-e2e317f79d6a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.610406] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49bb62b-93fa-4e3d-a749-9cd8d740f00c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.626490] env[65385]: DEBUG nova.compute.provider_tree [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1232.644843] env[65385]: DEBUG oslo_vmware.api [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Task: {'id': task-4454676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121355} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.645127] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1232.645306] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1232.645475] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1232.645637] env[65385]: INFO nova.compute.manager [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1232.645865] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1232.646066] env[65385]: DEBUG nova.compute.manager [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1232.646166] env[65385]: DEBUG nova.network.neutron [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1232.646402] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1232.646910] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1232.647183] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1232.685162] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1233.116045] env[65385]: DEBUG nova.compute.manager [req-837d773c-d2fd-49f7-9f67-2eccd0bd194f req-6ef8752b-ffec-4e92-b3fa-b57fe220c1c6 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Received event network-vif-deleted-cf3df048-713a-474d-9278-a467e50ec4f8 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1233.116255] env[65385]: INFO nova.compute.manager [req-837d773c-d2fd-49f7-9f67-2eccd0bd194f req-6ef8752b-ffec-4e92-b3fa-b57fe220c1c6 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Neutron deleted interface cf3df048-713a-474d-9278-a467e50ec4f8; detaching it from the instance and deleting it from the info cache [ 1233.116446] env[65385]: DEBUG nova.network.neutron [req-837d773c-d2fd-49f7-9f67-2eccd0bd194f req-6ef8752b-ffec-4e92-b3fa-b57fe220c1c6 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1233.129888] env[65385]: DEBUG nova.scheduler.client.report [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1233.135126] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1233.138160] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb6d0140-ceef-4b16-b958-dde3323f1d87 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.146911] env[65385]: DEBUG oslo_vmware.api [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1233.146911] env[65385]: value = "task-4454677" [ 1233.146911] env[65385]: _type = "Task" [ 1233.146911] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.157229] env[65385]: DEBUG oslo_vmware.api [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.593263] env[65385]: DEBUG nova.network.neutron [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1233.618649] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6fcb74e-54b2-4b69-9f21-ed32427a2a6f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.630105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cacf06fd-421c-4d65-90cf-c2b6a285c531 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.641854] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.653s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.656811] env[65385]: DEBUG oslo_vmware.api [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454677, 'name': PowerOffVM_Task, 'duration_secs': 0.170243} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.665302] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1233.665601] env[65385]: DEBUG nova.compute.manager [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1233.665955] env[65385]: DEBUG nova.compute.manager [req-837d773c-d2fd-49f7-9f67-2eccd0bd194f req-6ef8752b-ffec-4e92-b3fa-b57fe220c1c6 service nova] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Detach interface failed, port_id=cf3df048-713a-474d-9278-a467e50ec4f8, reason: Instance 20933973-82ae-498c-b016-3a82bb5dc165 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1233.667010] env[65385]: INFO nova.scheduler.client.report [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Deleted allocations for instance df419705-e7e8-47b4-b9a0-2f1cf2638f33 [ 1233.668608] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448bdd1f-241e-4097-94e5-8d0ffcef327a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.988985] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.989238] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1234.096046] env[65385]: INFO nova.compute.manager [-] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Took 1.45 seconds to deallocate network for instance. [ 1234.178826] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e3556715-1943-4ca5-b23e-bbde6a718264 tempest-AttachVolumeShelveTestJSON-37774609 tempest-AttachVolumeShelveTestJSON-37774609-project-member] Lock "df419705-e7e8-47b4-b9a0-2f1cf2638f33" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.833s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.184645] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a4696510-78e5-4d01-93af-4a1dd0b3eaff tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.078s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.491978] env[65385]: DEBUG nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1234.605653] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1234.606193] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1234.606517] env[65385]: DEBUG nova.objects.instance [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lazy-loading 'resources' on Instance uuid 20933973-82ae-498c-b016-3a82bb5dc165 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.609017] env[65385]: DEBUG nova.objects.instance [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'flavor' on Instance uuid fa02564b-c1a8-4f8f-876c-806618e37f5b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.016046] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.113938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.114145] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1235.114317] env[65385]: DEBUG nova.network.neutron [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1235.114481] env[65385]: DEBUG nova.objects.instance [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'info_cache' on Instance uuid fa02564b-c1a8-4f8f-876c-806618e37f5b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.170083] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee013c02-3e7a-42ba-9ef3-cf535bd49d7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.178717] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ed7449-5a59-483e-a730-f378690580f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.211330] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3226a85-6014-48c3-a558-aadc34f98d4e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.219598] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e222ab-95ef-4e1f-b769-01204946d122 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.234350] env[65385]: DEBUG nova.compute.provider_tree [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.618764] env[65385]: DEBUG nova.objects.base [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1235.738309] env[65385]: DEBUG nova.scheduler.client.report [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1236.121724] env[65385]: WARNING neutronclient.v2_0.client [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.122505] env[65385]: WARNING openstack [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.122854] env[65385]: WARNING openstack [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.243302] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.245780] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.230s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.247365] env[65385]: INFO nova.compute.claims [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1236.257559] env[65385]: WARNING openstack [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.257920] env[65385]: WARNING openstack [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.269788] env[65385]: INFO nova.scheduler.client.report [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Deleted allocations for instance 20933973-82ae-498c-b016-3a82bb5dc165 [ 1236.321031] env[65385]: WARNING neutronclient.v2_0.client [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.321718] env[65385]: WARNING openstack [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.322137] env[65385]: WARNING openstack [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.401669] env[65385]: DEBUG nova.network.neutron [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updating instance_info_cache with network_info: [{"id": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "address": "fa:16:3e:83:87:18", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d33bf4-66", "ovs_interfaceid": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.778041] env[65385]: DEBUG oslo_concurrency.lockutils [None req-0d4c694a-18f2-498f-8c9c-9d3587b6cc3a tempest-ServerActionsTestOtherB-140608432 tempest-ServerActionsTestOtherB-140608432-project-member] Lock "20933973-82ae-498c-b016-3a82bb5dc165" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.757s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.905464] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1237.311465] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93aa79da-3a85-40a2-9423-515fd7060710 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.321610] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999190ef-5019-4db9-abc1-b373cf6c23c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.353682] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2130d7fc-344b-4a59-ac50-c5c900fadb94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.362101] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef06510-e9c9-4917-9ce9-b6b4f03529f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.375743] env[65385]: DEBUG nova.compute.provider_tree [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.879223] env[65385]: DEBUG nova.scheduler.client.report [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1237.911274] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1237.911565] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-007344a5-9bc2-4895-9c23-dca5c52ece6b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.919564] env[65385]: DEBUG oslo_vmware.api [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1237.919564] env[65385]: value = "task-4454680" [ 1237.919564] env[65385]: _type = "Task" [ 1237.919564] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.928159] env[65385]: DEBUG oslo_vmware.api [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.385136] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.138s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.385136] env[65385]: DEBUG nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1238.430778] env[65385]: DEBUG oslo_vmware.api [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454680, 'name': PowerOnVM_Task, 'duration_secs': 0.410221} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.431601] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1238.431800] env[65385]: DEBUG nova.compute.manager [None req-9e934fc6-966b-4e40-af5a-0b01d1ee6c21 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1238.432612] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc9ee28-c007-4444-bc9c-dd0898fac44c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.844175] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1238.890298] env[65385]: DEBUG nova.compute.utils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1238.892346] env[65385]: DEBUG nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1238.892597] env[65385]: DEBUG nova.network.neutron [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1238.892930] env[65385]: WARNING neutronclient.v2_0.client [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1238.893264] env[65385]: WARNING neutronclient.v2_0.client [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1238.893869] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.894304] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.946389] env[65385]: DEBUG nova.policy [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b450c500f84f2fa682d70698c398b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d45ff3446a84ef2a1945f7a66b747e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1239.229057] env[65385]: DEBUG nova.network.neutron [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Successfully created port: 669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1239.402504] env[65385]: DEBUG nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1239.493134] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22444d9d-c3df-492d-8d64-f2133d2e2081 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.500153] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba63b8ca-8f3a-4551-9b38-eb20c71128c8 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Suspending the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1239.500427] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5dfc5e9b-bf52-4628-81d9-0472b890a447 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.508190] env[65385]: DEBUG oslo_vmware.api [None req-ba63b8ca-8f3a-4551-9b38-eb20c71128c8 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1239.508190] env[65385]: value = "task-4454681" [ 1239.508190] env[65385]: _type = "Task" [ 1239.508190] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.518588] env[65385]: DEBUG oslo_vmware.api [None req-ba63b8ca-8f3a-4551-9b38-eb20c71128c8 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454681, 'name': SuspendVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.844683] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.019713] env[65385]: DEBUG oslo_vmware.api [None req-ba63b8ca-8f3a-4551-9b38-eb20c71128c8 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454681, 'name': SuspendVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.412890] env[65385]: DEBUG nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1240.442273] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1240.442516] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1240.442706] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1240.442901] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1240.443474] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1240.443474] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1240.443474] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1240.443629] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1240.443725] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1240.443858] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1240.444104] env[65385]: DEBUG nova.virt.hardware [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1240.445107] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed76ddc-e1c4-443e-92bc-f4d9586f1186 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.454332] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44815ef8-115e-489c-bdf3-d3fe55786c3a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.519629] env[65385]: DEBUG oslo_vmware.api [None req-ba63b8ca-8f3a-4551-9b38-eb20c71128c8 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454681, 'name': SuspendVM_Task, 'duration_secs': 0.623382} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.519834] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba63b8ca-8f3a-4551-9b38-eb20c71128c8 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Suspended the VM {{(pid=65385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1240.520030] env[65385]: DEBUG nova.compute.manager [None req-ba63b8ca-8f3a-4551-9b38-eb20c71128c8 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1240.520838] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1372aed0-a98e-4276-ada3-c0f881ac9548 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.776165] env[65385]: DEBUG nova.network.neutron [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Successfully updated port: 669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1240.854017] env[65385]: DEBUG nova.compute.manager [req-e97b178c-f866-4ebf-83d6-57748fb38911 req-67945ca9-e05b-4fba-ab2b-70fb9c6e03bb service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Received event network-vif-plugged-669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1240.854017] env[65385]: DEBUG oslo_concurrency.lockutils [req-e97b178c-f866-4ebf-83d6-57748fb38911 req-67945ca9-e05b-4fba-ab2b-70fb9c6e03bb service nova] Acquiring lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1240.854017] env[65385]: DEBUG oslo_concurrency.lockutils [req-e97b178c-f866-4ebf-83d6-57748fb38911 req-67945ca9-e05b-4fba-ab2b-70fb9c6e03bb service nova] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.854017] env[65385]: DEBUG oslo_concurrency.lockutils [req-e97b178c-f866-4ebf-83d6-57748fb38911 req-67945ca9-e05b-4fba-ab2b-70fb9c6e03bb service nova] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.854017] env[65385]: DEBUG nova.compute.manager [req-e97b178c-f866-4ebf-83d6-57748fb38911 req-67945ca9-e05b-4fba-ab2b-70fb9c6e03bb service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] No waiting events found dispatching network-vif-plugged-669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1240.854769] env[65385]: WARNING nova.compute.manager [req-e97b178c-f866-4ebf-83d6-57748fb38911 req-67945ca9-e05b-4fba-ab2b-70fb9c6e03bb service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Received unexpected event network-vif-plugged-669739ea-2cff-42c7-a832-8b11965d1cdb for instance with vm_state building and task_state spawning. [ 1241.278682] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.279071] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1241.279569] env[65385]: DEBUG nova.network.neutron [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1241.782915] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.783384] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.818663] env[65385]: DEBUG nova.network.neutron [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1241.837969] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.838393] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.847808] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.848381] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.848429] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1241.848678] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.957939] env[65385]: WARNING neutronclient.v2_0.client [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.959247] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.959247] env[65385]: WARNING openstack [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.063061] env[65385]: DEBUG nova.network.neutron [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Updating instance_info_cache with network_info: [{"id": "669739ea-2cff-42c7-a832-8b11965d1cdb", "address": "fa:16:3e:a0:ac:fc", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap669739ea-2c", "ovs_interfaceid": "669739ea-2cff-42c7-a832-8b11965d1cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1242.352482] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.352798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.352845] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.353029] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1242.353910] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4960a856-070b-4d5e-8f37-cd23e664225d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.363102] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6d3490-d538-4e87-962b-9a18115d62a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.378302] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93716d67-2824-4265-b9d8-eba69eef2f00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.385062] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1896383-2b55-4fb5-88d6-45f05ba96fed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.418615] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179231MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1242.418615] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.418615] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.566409] env[65385]: INFO nova.compute.manager [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Resuming [ 1242.567127] env[65385]: DEBUG nova.objects.instance [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'flavor' on Instance uuid fa02564b-c1a8-4f8f-876c-806618e37f5b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1242.568671] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.569040] env[65385]: DEBUG nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Instance network_info: |[{"id": "669739ea-2cff-42c7-a832-8b11965d1cdb", "address": "fa:16:3e:a0:ac:fc", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap669739ea-2c", "ovs_interfaceid": "669739ea-2cff-42c7-a832-8b11965d1cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1242.570452] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:ac:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '669739ea-2cff-42c7-a832-8b11965d1cdb', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1242.577808] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1242.578350] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1242.578591] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a0b714e-e2c6-4aca-82f5-21910320d8cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.600359] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1242.600359] env[65385]: value = "task-4454682" [ 1242.600359] env[65385]: _type = "Task" [ 1242.600359] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.609796] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454682, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.902248] env[65385]: DEBUG nova.compute.manager [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Received event network-changed-669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1242.902459] env[65385]: DEBUG nova.compute.manager [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Refreshing instance network info cache due to event network-changed-669739ea-2cff-42c7-a832-8b11965d1cdb. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1242.902680] env[65385]: DEBUG oslo_concurrency.lockutils [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Acquiring lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.902825] env[65385]: DEBUG oslo_concurrency.lockutils [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Acquired lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1242.902983] env[65385]: DEBUG nova.network.neutron [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Refreshing network info cache for port 669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1243.110537] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454682, 'name': CreateVM_Task, 'duration_secs': 0.32874} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.110724] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1243.111216] env[65385]: WARNING neutronclient.v2_0.client [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.111573] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.111722] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.112044] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1243.112300] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-508dd603-a13f-4254-9bc1-51f1191b9b48 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.117477] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1243.117477] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52289c1c-9ed1-4405-17a2-ef402e1b6964" [ 1243.117477] env[65385]: _type = "Task" [ 1243.117477] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.126581] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52289c1c-9ed1-4405-17a2-ef402e1b6964, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.405737] env[65385]: WARNING neutronclient.v2_0.client [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.406531] env[65385]: WARNING openstack [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.406875] env[65385]: WARNING openstack [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.446012] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b4bca2c8-fc8f-471a-899b-aac648fee9fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1243.446306] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance fa02564b-c1a8-4f8f-876c-806618e37f5b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1243.446510] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance c9f09e05-52c4-44fb-bfa8-85bde23ef8ce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1243.446789] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1243.447054] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=100GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '3', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '3', 'num_proj_8d45ff3446a84ef2a1945f7a66b747e8': '2', 'io_workload': '1', 'num_vm_suspended': '1', 'num_task_resuming': '1', 'num_proj_60de0d1c162342209795a180391661f6': '1', 'num_vm_building': '1', 'num_task_spawning': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1243.503105] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85639810-ccc0-483d-911d-6ddffbe62270 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.517662] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b15107-ca6e-4b17-b00c-349e2221d830 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.555022] env[65385]: WARNING openstack [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.555454] env[65385]: WARNING openstack [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.562572] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2320e9d-2317-4500-abc0-b6fb3e3768e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.572025] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a21e7e-dcf6-4cb1-9f95-90effa9316ac {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.589756] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1243.592975] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.592975] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquired lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.593127] env[65385]: DEBUG nova.network.neutron [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1243.630391] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52289c1c-9ed1-4405-17a2-ef402e1b6964, 'name': SearchDatastore_Task, 'duration_secs': 0.012874} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.630697] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1243.630933] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1243.631186] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.631332] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.631505] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1243.631828] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a8d2dbd-9a64-497d-87c2-4c9f2901b09d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.642375] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1243.642583] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1243.643393] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1506c220-c06e-4891-95df-3494a42ff7fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.650254] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1243.650254] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d95793-4a66-fb3d-f843-f58fd06f821d" [ 1243.650254] env[65385]: _type = "Task" [ 1243.650254] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.658862] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d95793-4a66-fb3d-f843-f58fd06f821d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.660177] env[65385]: WARNING neutronclient.v2_0.client [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.660791] env[65385]: WARNING openstack [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.661162] env[65385]: WARNING openstack [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.745621] env[65385]: DEBUG nova.network.neutron [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Updated VIF entry in instance network info cache for port 669739ea-2cff-42c7-a832-8b11965d1cdb. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1243.746017] env[65385]: DEBUG nova.network.neutron [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Updating instance_info_cache with network_info: [{"id": "669739ea-2cff-42c7-a832-8b11965d1cdb", "address": "fa:16:3e:a0:ac:fc", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap669739ea-2c", "ovs_interfaceid": "669739ea-2cff-42c7-a832-8b11965d1cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1244.100755] env[65385]: WARNING neutronclient.v2_0.client [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.102499] env[65385]: WARNING openstack [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.102499] env[65385]: WARNING openstack [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.109911] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1244.161055] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52d95793-4a66-fb3d-f843-f58fd06f821d, 'name': SearchDatastore_Task, 'duration_secs': 0.014789} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.161739] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8b83f60-58c9-458e-b1c9-e3543a9df28b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.167661] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1244.167661] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c1530-1dff-ccfb-d5ce-8d499d5014ff" [ 1244.167661] env[65385]: _type = "Task" [ 1244.167661] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.175682] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c1530-1dff-ccfb-d5ce-8d499d5014ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.249152] env[65385]: DEBUG oslo_concurrency.lockutils [req-a19d5701-6cae-46c5-bc20-c8b29ccd6c21 req-2b7a2d7f-f85f-4c2e-99f1-e44d9e624816 service nova] Releasing lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.264645] env[65385]: WARNING openstack [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.265078] env[65385]: WARNING openstack [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.356702] env[65385]: WARNING neutronclient.v2_0.client [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.357397] env[65385]: WARNING openstack [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.357754] env[65385]: WARNING openstack [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.451236] env[65385]: DEBUG nova.network.neutron [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updating instance_info_cache with network_info: [{"id": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "address": "fa:16:3e:83:87:18", "network": {"id": "cf430c99-c961-4591-a0fb-fccdf963afde", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-917578284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60de0d1c162342209795a180391661f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d33bf4-66", "ovs_interfaceid": "34d33bf4-6698-4d35-a2ee-636a6e0fdbec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1244.614976] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1244.615297] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.197s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1244.681090] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c1530-1dff-ccfb-d5ce-8d499d5014ff, 'name': SearchDatastore_Task, 'duration_secs': 0.01058} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.681447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.681764] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c9f09e05-52c4-44fb-bfa8-85bde23ef8ce/c9f09e05-52c4-44fb-bfa8-85bde23ef8ce.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1244.682232] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8eefd062-3316-4207-9815-b162375f7f6f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.691685] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1244.691685] env[65385]: value = "task-4454683" [ 1244.691685] env[65385]: _type = "Task" [ 1244.691685] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.701559] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454683, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.955050] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Releasing lock "refresh_cache-fa02564b-c1a8-4f8f-876c-806618e37f5b" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.956183] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9794cb-7afc-4b50-98d2-38782bdead2e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.966641] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Resuming the VM {{(pid=65385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1244.967328] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1009920f-f510-44c6-827a-83da33f368e9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.975304] env[65385]: DEBUG oslo_vmware.api [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1244.975304] env[65385]: value = "task-4454684" [ 1244.975304] env[65385]: _type = "Task" [ 1244.975304] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.985718] env[65385]: DEBUG oslo_vmware.api [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454684, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.204191] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454683, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.486225] env[65385]: DEBUG oslo_vmware.api [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454684, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.706074] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454683, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615539} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.706074] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] c9f09e05-52c4-44fb-bfa8-85bde23ef8ce/c9f09e05-52c4-44fb-bfa8-85bde23ef8ce.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1245.706074] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1245.706332] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4e2d07d-69be-4a54-a5b3-9ed5b35077b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.714467] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1245.714467] env[65385]: value = "task-4454685" [ 1245.714467] env[65385]: _type = "Task" [ 1245.714467] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.726064] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454685, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.987338] env[65385]: DEBUG oslo_vmware.api [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454684, 'name': PowerOnVM_Task} progress is 93%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.226994] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454685, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.285368} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.227345] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1246.228211] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df66e99b-59ca-4473-abb5-11b39fbc4ed5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.253507] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] c9f09e05-52c4-44fb-bfa8-85bde23ef8ce/c9f09e05-52c4-44fb-bfa8-85bde23ef8ce.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.253876] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-456ec28b-baf7-466f-8fc1-8f040619ea0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.277448] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1246.277448] env[65385]: value = "task-4454686" [ 1246.277448] env[65385]: _type = "Task" [ 1246.277448] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.290396] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454686, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.404143] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "120aa07c-d0c1-4031-8b61-986a261f2291" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.404386] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "120aa07c-d0c1-4031-8b61-986a261f2291" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.487965] env[65385]: DEBUG oslo_vmware.api [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454684, 'name': PowerOnVM_Task, 'duration_secs': 1.097502} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.488278] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Resumed the VM {{(pid=65385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1246.488462] env[65385]: DEBUG nova.compute.manager [None req-8b7dc1a8-b78e-417a-a54a-c6fe5adfb351 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1246.489329] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d2797a-d8c0-4660-ae2c-c73ea5ef40ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.611803] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.611803] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.790499] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454686, 'name': ReconfigVM_Task, 'duration_secs': 0.501611} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.790788] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Reconfigured VM instance instance-00000077 to attach disk [datastore2] c9f09e05-52c4-44fb-bfa8-85bde23ef8ce/c9f09e05-52c4-44fb-bfa8-85bde23ef8ce.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1246.791479] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f438bea-6174-4895-8344-18691190dd9b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.799871] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1246.799871] env[65385]: value = "task-4454687" [ 1246.799871] env[65385]: _type = "Task" [ 1246.799871] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.812063] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454687, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.839206] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.907515] env[65385]: DEBUG nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1247.311734] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454687, 'name': Rename_Task, 'duration_secs': 0.160989} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.312056] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1247.312455] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c385bdc9-0fb0-492c-bfe6-92c73f62550f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.321196] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1247.321196] env[65385]: value = "task-4454688" [ 1247.321196] env[65385]: _type = "Task" [ 1247.321196] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.330405] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454688, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.344623] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.367490] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1247.367769] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1247.367979] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "fa02564b-c1a8-4f8f-876c-806618e37f5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1247.368189] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1247.368362] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1247.370621] env[65385]: INFO nova.compute.manager [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Terminating instance [ 1247.432773] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1247.433059] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1247.434648] env[65385]: INFO nova.compute.claims [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1247.831684] env[65385]: DEBUG oslo_vmware.api [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454688, 'name': PowerOnVM_Task, 'duration_secs': 0.475627} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.832032] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1247.832225] env[65385]: INFO nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Took 7.42 seconds to spawn the instance on the hypervisor. [ 1247.832405] env[65385]: DEBUG nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1247.833222] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6cda0e-6ac8-4548-a396-9a0478d2d977 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.876578] env[65385]: DEBUG nova.compute.manager [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1247.876836] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1247.877749] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5ffa9c-7f25-4b06-b4c1-ce660afcd73d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.886812] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1247.887019] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15144271-be8f-4b2b-a77c-12fb54c38147 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.894593] env[65385]: DEBUG oslo_vmware.api [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1247.894593] env[65385]: value = "task-4454689" [ 1247.894593] env[65385]: _type = "Task" [ 1247.894593] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.904181] env[65385]: DEBUG oslo_vmware.api [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.351302] env[65385]: INFO nova.compute.manager [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Took 13.35 seconds to build instance. [ 1248.405855] env[65385]: DEBUG oslo_vmware.api [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454689, 'name': PowerOffVM_Task, 'duration_secs': 0.215804} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.406134] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1248.406307] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1248.406564] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16468801-0f71-4283-931a-a64e18c267bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.479225] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1248.479417] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1248.479596] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleting the datastore file [datastore1] fa02564b-c1a8-4f8f-876c-806618e37f5b {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1248.480090] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-627524a6-6ff3-4cff-b47b-3f23531b337b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.486318] env[65385]: DEBUG oslo_vmware.api [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for the task: (returnval){ [ 1248.486318] env[65385]: value = "task-4454691" [ 1248.486318] env[65385]: _type = "Task" [ 1248.486318] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.497990] env[65385]: DEBUG oslo_vmware.api [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454691, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.514662] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75ccac6-8abd-4c94-a70e-6eb1a2c71803 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.522400] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0e2b1b-b9c2-449a-b44b-75fd37476b6a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.552009] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26eb225d-ad15-4e9e-a9a1-c2e1f3cc6f4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.559881] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2c4456-2f93-469d-b265-6e26b7c6f09e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.573546] env[65385]: DEBUG nova.compute.provider_tree [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.853640] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bcdc93a6-88c8-45a3-8bed-8c485e64047d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.864s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.997828] env[65385]: DEBUG oslo_vmware.api [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Task: {'id': task-4454691, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157764} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.998115] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1248.998305] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1248.998488] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1248.998658] env[65385]: INFO nova.compute.manager [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1248.998912] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1248.999124] env[65385]: DEBUG nova.compute.manager [-] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1248.999228] env[65385]: DEBUG nova.network.neutron [-] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1248.999489] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1249.000014] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1249.000281] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1249.037423] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1249.077466] env[65385]: DEBUG nova.scheduler.client.report [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1249.246282] env[65385]: DEBUG nova.compute.manager [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Received event network-changed-669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1249.246282] env[65385]: DEBUG nova.compute.manager [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Refreshing instance network info cache due to event network-changed-669739ea-2cff-42c7-a832-8b11965d1cdb. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1249.246282] env[65385]: DEBUG oslo_concurrency.lockutils [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Acquiring lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1249.246515] env[65385]: DEBUG oslo_concurrency.lockutils [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Acquired lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1249.246565] env[65385]: DEBUG nova.network.neutron [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Refreshing network info cache for port 669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1249.462782] env[65385]: DEBUG nova.compute.manager [req-8a1502bb-aeb0-436a-a8bb-8b402386979b req-41a2d628-0dd4-4b90-aaf2-900674d005df service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Received event network-vif-deleted-34d33bf4-6698-4d35-a2ee-636a6e0fdbec {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1249.462782] env[65385]: INFO nova.compute.manager [req-8a1502bb-aeb0-436a-a8bb-8b402386979b req-41a2d628-0dd4-4b90-aaf2-900674d005df service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Neutron deleted interface 34d33bf4-6698-4d35-a2ee-636a6e0fdbec; detaching it from the instance and deleting it from the info cache [ 1249.462948] env[65385]: DEBUG nova.network.neutron [req-8a1502bb-aeb0-436a-a8bb-8b402386979b req-41a2d628-0dd4-4b90-aaf2-900674d005df service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1249.583310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.150s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1249.584483] env[65385]: DEBUG nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1249.753561] env[65385]: WARNING neutronclient.v2_0.client [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1249.754342] env[65385]: WARNING openstack [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1249.754680] env[65385]: WARNING openstack [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1249.863997] env[65385]: WARNING openstack [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1249.864487] env[65385]: WARNING openstack [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1249.928680] env[65385]: WARNING neutronclient.v2_0.client [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1249.928844] env[65385]: WARNING openstack [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1249.929405] env[65385]: WARNING openstack [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1249.944080] env[65385]: DEBUG nova.network.neutron [-] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1249.965298] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e5b8347-c62c-495b-9942-9221b28eef8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.980722] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d860782-69f7-4bf4-bcda-b50d4f5a9de9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.016657] env[65385]: DEBUG nova.compute.manager [req-8a1502bb-aeb0-436a-a8bb-8b402386979b req-41a2d628-0dd4-4b90-aaf2-900674d005df service nova] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Detach interface failed, port_id=34d33bf4-6698-4d35-a2ee-636a6e0fdbec, reason: Instance fa02564b-c1a8-4f8f-876c-806618e37f5b could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1250.089966] env[65385]: DEBUG nova.compute.utils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1250.091478] env[65385]: DEBUG nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1250.091678] env[65385]: DEBUG nova.network.neutron [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1250.092046] env[65385]: WARNING neutronclient.v2_0.client [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.092438] env[65385]: WARNING neutronclient.v2_0.client [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.093054] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1250.093419] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1250.163133] env[65385]: DEBUG nova.policy [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '043f66592e52480c87c36a10799be03c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96214fc566ed4b11a92a6369a3419aea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1250.168046] env[65385]: DEBUG nova.network.neutron [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Updated VIF entry in instance network info cache for port 669739ea-2cff-42c7-a832-8b11965d1cdb. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1250.168583] env[65385]: DEBUG nova.network.neutron [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Updating instance_info_cache with network_info: [{"id": "669739ea-2cff-42c7-a832-8b11965d1cdb", "address": "fa:16:3e:a0:ac:fc", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap669739ea-2c", "ovs_interfaceid": "669739ea-2cff-42c7-a832-8b11965d1cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1250.448042] env[65385]: INFO nova.compute.manager [-] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Took 1.45 seconds to deallocate network for instance. [ 1250.506343] env[65385]: DEBUG nova.network.neutron [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Successfully created port: 0fcb3c6e-b016-42ef-aec2-92569a585155 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1250.601651] env[65385]: DEBUG nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1250.671458] env[65385]: DEBUG oslo_concurrency.lockutils [req-294daaee-b71a-478d-ab8a-322823371237 req-89eb7803-48c0-437c-afe9-4069c20045b5 service nova] Releasing lock "refresh_cache-c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1250.958938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.959285] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.959497] env[65385]: DEBUG nova.objects.instance [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lazy-loading 'resources' on Instance uuid fa02564b-c1a8-4f8f-876c-806618e37f5b {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1251.534881] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b64909f-818c-4bf2-87a9-838bd7384598 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.543247] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de14830-0828-4c63-a3ae-7e057cdd89b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.572972] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb2814e-1240-42e5-a77c-16813bfedd9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.580829] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7ade57-28cb-468c-9170-e93986200e31 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.594556] env[65385]: DEBUG nova.compute.provider_tree [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.612315] env[65385]: DEBUG nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1251.639524] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1251.639796] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1251.639979] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1251.640180] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1251.640350] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1251.640505] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1251.640714] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1251.640869] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1251.641039] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1251.641223] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1251.641427] env[65385]: DEBUG nova.virt.hardware [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1251.642274] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbaac2e6-8e6e-4df7-98f5-ae57ffc05e92 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.651156] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a75cda1-bb30-4068-8086-cfee6fd9cc34 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.977498] env[65385]: DEBUG nova.compute.manager [req-f4c12bef-8cd0-4180-8ae8-ec3cc86c5ca2 req-17a11389-3c69-4186-b0e6-c8ac72a603ff service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Received event network-vif-plugged-0fcb3c6e-b016-42ef-aec2-92569a585155 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1251.977750] env[65385]: DEBUG oslo_concurrency.lockutils [req-f4c12bef-8cd0-4180-8ae8-ec3cc86c5ca2 req-17a11389-3c69-4186-b0e6-c8ac72a603ff service nova] Acquiring lock "120aa07c-d0c1-4031-8b61-986a261f2291-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1251.977856] env[65385]: DEBUG oslo_concurrency.lockutils [req-f4c12bef-8cd0-4180-8ae8-ec3cc86c5ca2 req-17a11389-3c69-4186-b0e6-c8ac72a603ff service nova] Lock "120aa07c-d0c1-4031-8b61-986a261f2291-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1251.978081] env[65385]: DEBUG oslo_concurrency.lockutils [req-f4c12bef-8cd0-4180-8ae8-ec3cc86c5ca2 req-17a11389-3c69-4186-b0e6-c8ac72a603ff service nova] Lock "120aa07c-d0c1-4031-8b61-986a261f2291-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1251.978249] env[65385]: DEBUG nova.compute.manager [req-f4c12bef-8cd0-4180-8ae8-ec3cc86c5ca2 req-17a11389-3c69-4186-b0e6-c8ac72a603ff service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] No waiting events found dispatching network-vif-plugged-0fcb3c6e-b016-42ef-aec2-92569a585155 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1251.978407] env[65385]: WARNING nova.compute.manager [req-f4c12bef-8cd0-4180-8ae8-ec3cc86c5ca2 req-17a11389-3c69-4186-b0e6-c8ac72a603ff service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Received unexpected event network-vif-plugged-0fcb3c6e-b016-42ef-aec2-92569a585155 for instance with vm_state building and task_state spawning. [ 1252.057747] env[65385]: DEBUG nova.network.neutron [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Successfully updated port: 0fcb3c6e-b016-42ef-aec2-92569a585155 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1252.097368] env[65385]: DEBUG nova.scheduler.client.report [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1252.561206] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.561459] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1252.562569] env[65385]: DEBUG nova.network.neutron [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1252.601759] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.642s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1252.632633] env[65385]: INFO nova.scheduler.client.report [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Deleted allocations for instance fa02564b-c1a8-4f8f-876c-806618e37f5b [ 1253.065294] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1253.065930] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1253.120263] env[65385]: DEBUG nova.network.neutron [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1253.141667] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1253.142227] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1253.150164] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ba7aee39-6cd3-483e-9d98-5e62b796b732 tempest-ServerActionsTestJSON-849609147 tempest-ServerActionsTestJSON-849609147-project-member] Lock "fa02564b-c1a8-4f8f-876c-806618e37f5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.782s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.209469] env[65385]: WARNING neutronclient.v2_0.client [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1253.210313] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1253.210830] env[65385]: WARNING openstack [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1253.306240] env[65385]: DEBUG nova.network.neutron [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Updating instance_info_cache with network_info: [{"id": "0fcb3c6e-b016-42ef-aec2-92569a585155", "address": "fa:16:3e:fe:6e:2e", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fcb3c6e-b0", "ovs_interfaceid": "0fcb3c6e-b016-42ef-aec2-92569a585155", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1253.809166] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1253.809567] env[65385]: DEBUG nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Instance network_info: |[{"id": "0fcb3c6e-b016-42ef-aec2-92569a585155", "address": "fa:16:3e:fe:6e:2e", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fcb3c6e-b0", "ovs_interfaceid": "0fcb3c6e-b016-42ef-aec2-92569a585155", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1253.810048] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:6e:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe99da4f-5630-4afd-918b-b327193d8489', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0fcb3c6e-b016-42ef-aec2-92569a585155', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1253.817611] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Creating folder: Project (96214fc566ed4b11a92a6369a3419aea). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1253.817949] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44b3e29c-591d-4c62-96c6-6d53f40b14fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.830649] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Created folder: Project (96214fc566ed4b11a92a6369a3419aea) in parent group-v870881. [ 1253.830887] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Creating folder: Instances. Parent ref: group-v871208. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1253.831117] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cce48c7c-eadc-4157-9e44-95becd41173a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.841691] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Created folder: Instances in parent group-v871208. [ 1253.841923] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1253.842188] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1253.842412] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e810a0a4-cf5b-4784-a250-395c8889a484 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.863478] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1253.863478] env[65385]: value = "task-4454696" [ 1253.863478] env[65385]: _type = "Task" [ 1253.863478] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.872106] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454696, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.017417] env[65385]: DEBUG nova.compute.manager [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Received event network-changed-0fcb3c6e-b016-42ef-aec2-92569a585155 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1254.017652] env[65385]: DEBUG nova.compute.manager [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Refreshing instance network info cache due to event network-changed-0fcb3c6e-b016-42ef-aec2-92569a585155. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1254.017762] env[65385]: DEBUG oslo_concurrency.lockutils [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Acquiring lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.017913] env[65385]: DEBUG oslo_concurrency.lockutils [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Acquired lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1254.018255] env[65385]: DEBUG nova.network.neutron [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Refreshing network info cache for port 0fcb3c6e-b016-42ef-aec2-92569a585155 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1254.374620] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454696, 'name': CreateVM_Task, 'duration_secs': 0.379557} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.374967] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1254.376251] env[65385]: WARNING neutronclient.v2_0.client [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1254.376251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.376251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1254.376251] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1254.376469] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b72fecd-1fba-490b-a7f8-30d58557fa6e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.382390] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1254.382390] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527d00c6-5241-13d7-5f1d-aa637f05f710" [ 1254.382390] env[65385]: _type = "Task" [ 1254.382390] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.391386] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527d00c6-5241-13d7-5f1d-aa637f05f710, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.521355] env[65385]: WARNING neutronclient.v2_0.client [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1254.522017] env[65385]: WARNING openstack [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1254.522410] env[65385]: WARNING openstack [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1254.661807] env[65385]: WARNING openstack [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1254.662375] env[65385]: WARNING openstack [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1254.715957] env[65385]: WARNING neutronclient.v2_0.client [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1254.716753] env[65385]: WARNING openstack [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1254.717124] env[65385]: WARNING openstack [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1254.796386] env[65385]: DEBUG nova.network.neutron [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Updated VIF entry in instance network info cache for port 0fcb3c6e-b016-42ef-aec2-92569a585155. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1254.796753] env[65385]: DEBUG nova.network.neutron [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Updating instance_info_cache with network_info: [{"id": "0fcb3c6e-b016-42ef-aec2-92569a585155", "address": "fa:16:3e:fe:6e:2e", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fcb3c6e-b0", "ovs_interfaceid": "0fcb3c6e-b016-42ef-aec2-92569a585155", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1254.894670] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]527d00c6-5241-13d7-5f1d-aa637f05f710, 'name': SearchDatastore_Task, 'duration_secs': 0.009748} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.894670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1254.894670] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1254.894670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.894670] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1254.894973] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1254.894973] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c0ff945-5a5c-4780-9421-476291a16c1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.904765] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1254.904765] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1254.905345] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6de24d23-b0f1-45e8-bfa0-b3cfe2a995c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.911701] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1254.911701] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5264ce26-5039-5ad0-db4b-4fb6f5997c90" [ 1254.911701] env[65385]: _type = "Task" [ 1254.911701] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.920455] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5264ce26-5039-5ad0-db4b-4fb6f5997c90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.301034] env[65385]: DEBUG oslo_concurrency.lockutils [req-11546246-2cfe-457d-a9c7-7b51ef4b96dd req-5e146b51-95ae-42ea-bc36-c009945a594b service nova] Releasing lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1255.424886] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5264ce26-5039-5ad0-db4b-4fb6f5997c90, 'name': SearchDatastore_Task, 'duration_secs': 0.008823} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.426175] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-599dbdf5-2e71-46a8-a809-143ac212858f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.433727] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1255.433727] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520af5de-86be-bc7c-72f8-918112f11dbf" [ 1255.433727] env[65385]: _type = "Task" [ 1255.433727] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.443618] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520af5de-86be-bc7c-72f8-918112f11dbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.946821] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520af5de-86be-bc7c-72f8-918112f11dbf, 'name': SearchDatastore_Task, 'duration_secs': 0.01094} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.947120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1255.947344] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/120aa07c-d0c1-4031-8b61-986a261f2291.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1255.947644] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c61e188-6c7b-4888-a6d1-daae10cecca1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.955384] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1255.955384] env[65385]: value = "task-4454698" [ 1255.955384] env[65385]: _type = "Task" [ 1255.955384] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.964854] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.466082] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485365} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.466458] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/120aa07c-d0c1-4031-8b61-986a261f2291.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1256.466600] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1256.466854] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-adc5e83f-b886-479e-b07f-9efebdab96b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.474703] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1256.474703] env[65385]: value = "task-4454699" [ 1256.474703] env[65385]: _type = "Task" [ 1256.474703] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.482751] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454699, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.986016] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454699, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0627} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.986462] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1256.987067] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a9de44-5d23-478a-8bd4-83526ba4872d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.011180] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/120aa07c-d0c1-4031-8b61-986a261f2291.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1257.011402] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-145b7d10-f2c5-48ee-891a-1cf6ec2458d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.033301] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1257.033301] env[65385]: value = "task-4454700" [ 1257.033301] env[65385]: _type = "Task" [ 1257.033301] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.045022] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454700, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.548931] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454700, 'name': ReconfigVM_Task, 'duration_secs': 0.28188} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.551419] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Reconfigured VM instance instance-00000078 to attach disk [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/120aa07c-d0c1-4031-8b61-986a261f2291.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1257.551419] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9604b5b5-4f3b-44d6-b88b-1f40d1a9c511 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.558291] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1257.558291] env[65385]: value = "task-4454702" [ 1257.558291] env[65385]: _type = "Task" [ 1257.558291] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.572529] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454702, 'name': Rename_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.068629] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454702, 'name': Rename_Task, 'duration_secs': 0.147729} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.068939] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1258.069222] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8a40e4e-73e2-4ff4-a6ef-97c0c9e839d3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.076758] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1258.076758] env[65385]: value = "task-4454703" [ 1258.076758] env[65385]: _type = "Task" [ 1258.076758] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.085958] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454703, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.589066] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454703, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.089032] env[65385]: DEBUG oslo_vmware.api [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454703, 'name': PowerOnVM_Task, 'duration_secs': 0.589967} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.089344] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1259.089579] env[65385]: INFO nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1259.089861] env[65385]: DEBUG nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1259.090574] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221950fc-c3d4-4420-8769-64207288c598 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.612936] env[65385]: INFO nova.compute.manager [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Took 12.20 seconds to build instance. [ 1259.775738] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1259.775968] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1260.115054] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa8e77c4-7beb-4585-b6a2-a50b49237778 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "120aa07c-d0c1-4031-8b61-986a261f2291" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.710s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1260.280340] env[65385]: DEBUG nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1260.803358] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1260.803693] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1260.805329] env[65385]: INFO nova.compute.claims [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1260.941422] env[65385]: INFO nova.compute.manager [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Rescuing [ 1260.941721] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1260.941888] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1260.942066] env[65385]: DEBUG nova.network.neutron [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1261.445144] env[65385]: WARNING neutronclient.v2_0.client [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1261.445866] env[65385]: WARNING openstack [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.446244] env[65385]: WARNING openstack [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1261.557555] env[65385]: WARNING openstack [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.558019] env[65385]: WARNING openstack [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1261.606746] env[65385]: WARNING neutronclient.v2_0.client [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1261.607480] env[65385]: WARNING openstack [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.607856] env[65385]: WARNING openstack [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1261.688700] env[65385]: DEBUG nova.network.neutron [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Updating instance_info_cache with network_info: [{"id": "0fcb3c6e-b016-42ef-aec2-92569a585155", "address": "fa:16:3e:fe:6e:2e", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0fcb3c6e-b0", "ovs_interfaceid": "0fcb3c6e-b016-42ef-aec2-92569a585155", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1261.879763] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b80d7a-a668-470d-9871-ebdf4e46e8c2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.887960] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78483245-82d4-45f4-9c4c-b5242e0884dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.918292] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89d722b-ca5e-4ed4-8c33-c9771b5ce2a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.926041] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180ff2dd-90b9-4afa-ada6-bce24f0dc93b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.939466] env[65385]: DEBUG nova.compute.provider_tree [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1262.191622] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "refresh_cache-120aa07c-d0c1-4031-8b61-986a261f2291" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1262.443384] env[65385]: DEBUG nova.scheduler.client.report [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1262.948084] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.144s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1262.948606] env[65385]: DEBUG nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1263.454022] env[65385]: DEBUG nova.compute.utils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1263.455622] env[65385]: DEBUG nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1263.455836] env[65385]: DEBUG nova.network.neutron [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1263.456205] env[65385]: WARNING neutronclient.v2_0.client [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.456539] env[65385]: WARNING neutronclient.v2_0.client [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.457177] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.457570] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.517117] env[65385]: DEBUG nova.policy [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1394af239e8482d8efd42d6cdc96b5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df347ee234f142899014703c70ee42ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1263.727651] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1263.727989] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58eecaf1-26b6-4500-8b29-37b7e082a346 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.736045] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1263.736045] env[65385]: value = "task-4454707" [ 1263.736045] env[65385]: _type = "Task" [ 1263.736045] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.745083] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.815577] env[65385]: DEBUG nova.network.neutron [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Successfully created port: 59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1263.966726] env[65385]: DEBUG nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1264.246512] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454707, 'name': PowerOffVM_Task, 'duration_secs': 0.168962} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.246842] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1264.247676] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae978766-72fb-4748-982e-94a92f10e137 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.267236] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4366306f-77d0-4630-8560-8e36bd7ec146 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.303631] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1264.303942] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46e23097-a09d-4275-bcef-24bd21341db1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.312315] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1264.312315] env[65385]: value = "task-4454708" [ 1264.312315] env[65385]: _type = "Task" [ 1264.312315] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.320920] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454708, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.823569] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1264.823793] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1264.824050] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.824232] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1264.824418] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1264.824675] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b60630b5-3471-4888-bbe2-4b6b50567efb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.834313] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1264.834505] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1264.835274] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9808e40-5fbd-4dbc-b5e5-8a0b86428c2b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.841221] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1264.841221] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cba9aa-54c3-3db0-1954-ca98a516be0d" [ 1264.841221] env[65385]: _type = "Task" [ 1264.841221] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.852663] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52cba9aa-54c3-3db0-1954-ca98a516be0d, 'name': SearchDatastore_Task, 'duration_secs': 0.009038} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.853421] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bd887e8-265d-446b-a4e5-3f534db1e2b8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.858865] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1264.858865] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5230d673-6b6c-0906-5dc7-a2011cc703d8" [ 1264.858865] env[65385]: _type = "Task" [ 1264.858865] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.867197] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5230d673-6b6c-0906-5dc7-a2011cc703d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.978751] env[65385]: DEBUG nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1265.005359] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1265.005602] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1265.005754] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1265.005928] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1265.006116] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1265.006268] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1265.006483] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1265.006649] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1265.006813] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1265.006974] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1265.007158] env[65385]: DEBUG nova.virt.hardware [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1265.007996] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ac9252-67c4-46c1-a60a-09763ad9caa8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.016382] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d19b54-7e77-431c-b0b7-cd194d8b5100 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.208920] env[65385]: DEBUG nova.compute.manager [req-84deb080-3109-4841-a844-457827eeddd7 req-ab7d1a25-724c-4343-8fcc-80c5a7149311 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Received event network-vif-plugged-59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1265.209163] env[65385]: DEBUG oslo_concurrency.lockutils [req-84deb080-3109-4841-a844-457827eeddd7 req-ab7d1a25-724c-4343-8fcc-80c5a7149311 service nova] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1265.209424] env[65385]: DEBUG oslo_concurrency.lockutils [req-84deb080-3109-4841-a844-457827eeddd7 req-ab7d1a25-724c-4343-8fcc-80c5a7149311 service nova] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1265.209603] env[65385]: DEBUG oslo_concurrency.lockutils [req-84deb080-3109-4841-a844-457827eeddd7 req-ab7d1a25-724c-4343-8fcc-80c5a7149311 service nova] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1265.209771] env[65385]: DEBUG nova.compute.manager [req-84deb080-3109-4841-a844-457827eeddd7 req-ab7d1a25-724c-4343-8fcc-80c5a7149311 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] No waiting events found dispatching network-vif-plugged-59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1265.210019] env[65385]: WARNING nova.compute.manager [req-84deb080-3109-4841-a844-457827eeddd7 req-ab7d1a25-724c-4343-8fcc-80c5a7149311 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Received unexpected event network-vif-plugged-59e7b718-2455-470d-85a3-3a0d0849ea44 for instance with vm_state building and task_state spawning. [ 1265.288060] env[65385]: DEBUG nova.network.neutron [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Successfully updated port: 59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1265.371395] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5230d673-6b6c-0906-5dc7-a2011cc703d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009064} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.371655] env[65385]: DEBUG oslo_concurrency.lockutils [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1265.371915] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. {{(pid=65385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1265.372240] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87362e90-8067-497e-8b51-3b1264bede4f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.381216] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1265.381216] env[65385]: value = "task-4454709" [ 1265.381216] env[65385]: _type = "Task" [ 1265.381216] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.389694] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.791524] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.791771] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1265.792013] env[65385]: DEBUG nova.network.neutron [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1265.890734] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465387} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.890989] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. [ 1265.891779] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9498321c-80ea-4c32-b919-771001eda358 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.917922] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1265.918690] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e94f73c-ab4f-42e6-9398-28a250d1cc58 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.939021] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1265.939021] env[65385]: value = "task-4454711" [ 1265.939021] env[65385]: _type = "Task" [ 1265.939021] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.948317] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454711, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.294988] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1266.295413] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1266.331761] env[65385]: DEBUG nova.network.neutron [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1266.356205] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1266.356942] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1266.420729] env[65385]: WARNING neutronclient.v2_0.client [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1266.421416] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1266.421780] env[65385]: WARNING openstack [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1266.450528] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454711, 'name': ReconfigVM_Task, 'duration_secs': 0.295878} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.450793] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Reconfigured VM instance instance-00000078 to attach disk [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1266.451615] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28150dd0-742e-4735-947e-c812672cb9d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.480831] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f89fea87-495d-448d-8d7a-7b2e7119e58d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.498471] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1266.498471] env[65385]: value = "task-4454712" [ 1266.498471] env[65385]: _type = "Task" [ 1266.498471] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.507581] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454712, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.528084] env[65385]: DEBUG nova.network.neutron [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating instance_info_cache with network_info: [{"id": "59e7b718-2455-470d-85a3-3a0d0849ea44", "address": "fa:16:3e:7e:65:d9", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e7b718-24", "ovs_interfaceid": "59e7b718-2455-470d-85a3-3a0d0849ea44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1267.010032] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454712, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.031067] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1267.031381] env[65385]: DEBUG nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Instance network_info: |[{"id": "59e7b718-2455-470d-85a3-3a0d0849ea44", "address": "fa:16:3e:7e:65:d9", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e7b718-24", "ovs_interfaceid": "59e7b718-2455-470d-85a3-3a0d0849ea44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1267.031840] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:65:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ebd8af-aaf6-4d04-b869-3882e2571ed7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59e7b718-2455-470d-85a3-3a0d0849ea44', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1267.039385] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Creating folder: Project (df347ee234f142899014703c70ee42ed). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1267.039655] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3423b1d6-b7cb-44a2-bf37-bd48a2016a70 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.052627] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Created folder: Project (df347ee234f142899014703c70ee42ed) in parent group-v870881. [ 1267.053058] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Creating folder: Instances. Parent ref: group-v871212. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1267.053433] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b67b4a5-3054-4e9d-a00e-65686f1db661 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.065340] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Created folder: Instances in parent group-v871212. [ 1267.065573] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1267.065775] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1267.065993] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a4113a1-6322-4bec-b497-a44e91d8bfee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.085909] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1267.085909] env[65385]: value = "task-4454715" [ 1267.085909] env[65385]: _type = "Task" [ 1267.085909] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.094467] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454715, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.239795] env[65385]: DEBUG nova.compute.manager [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Received event network-changed-59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1267.240061] env[65385]: DEBUG nova.compute.manager [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Refreshing instance network info cache due to event network-changed-59e7b718-2455-470d-85a3-3a0d0849ea44. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1267.240277] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Acquiring lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1267.240414] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Acquired lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1267.240586] env[65385]: DEBUG nova.network.neutron [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Refreshing network info cache for port 59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1267.511700] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454712, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.597022] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454715, 'name': CreateVM_Task, 'duration_secs': 0.31365} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.597225] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1267.597806] env[65385]: WARNING neutronclient.v2_0.client [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.598216] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1267.598404] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1267.598744] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1267.599029] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84fb241b-0b32-46bb-9d14-1755af50b27b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.604303] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1267.604303] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f95ac2-264f-54b4-1782-df06dbc84d40" [ 1267.604303] env[65385]: _type = "Task" [ 1267.604303] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.613136] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f95ac2-264f-54b4-1782-df06dbc84d40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.743096] env[65385]: WARNING neutronclient.v2_0.client [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.743810] env[65385]: WARNING openstack [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.744253] env[65385]: WARNING openstack [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.853967] env[65385]: WARNING openstack [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.854477] env[65385]: WARNING openstack [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.922755] env[65385]: WARNING neutronclient.v2_0.client [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.923441] env[65385]: WARNING openstack [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.923800] env[65385]: WARNING openstack [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1268.006800] env[65385]: DEBUG nova.network.neutron [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updated VIF entry in instance network info cache for port 59e7b718-2455-470d-85a3-3a0d0849ea44. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1268.007237] env[65385]: DEBUG nova.network.neutron [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating instance_info_cache with network_info: [{"id": "59e7b718-2455-470d-85a3-3a0d0849ea44", "address": "fa:16:3e:7e:65:d9", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e7b718-24", "ovs_interfaceid": "59e7b718-2455-470d-85a3-3a0d0849ea44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1268.014221] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454712, 'name': ReconfigVM_Task, 'duration_secs': 1.182653} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.014484] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1268.014732] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-645a6ab5-939f-46b7-b39d-b527376374f5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.022113] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1268.022113] env[65385]: value = "task-4454716" [ 1268.022113] env[65385]: _type = "Task" [ 1268.022113] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.032804] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.115778] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f95ac2-264f-54b4-1782-df06dbc84d40, 'name': SearchDatastore_Task, 'duration_secs': 0.011234} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.116109] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1268.116351] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1268.116597] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1268.116736] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1268.116911] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1268.117190] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b244e2db-a0cd-4c4a-9899-a673b1b62a72 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.126200] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1268.126411] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1268.127155] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fa21d3a-cd52-49ea-bbf0-00bbb211ec79 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.132719] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1268.132719] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5267f578-112e-86b6-3e07-fb4ce08852c0" [ 1268.132719] env[65385]: _type = "Task" [ 1268.132719] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.141199] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5267f578-112e-86b6-3e07-fb4ce08852c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.510650] env[65385]: DEBUG oslo_concurrency.lockutils [req-3a49113f-c6e3-4200-8c4d-0d9a4e18e4be req-043666e0-a7ab-42ff-b8f7-7d9c2b20f742 service nova] Releasing lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1268.532672] env[65385]: DEBUG oslo_vmware.api [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454716, 'name': PowerOnVM_Task, 'duration_secs': 0.389072} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.532980] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1268.535865] env[65385]: DEBUG nova.compute.manager [None req-1fc2fcdf-5f28-4ef5-9ca1-42d05e1b6dea tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1268.537036] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d02ca9-5cc4-4378-a1d5-43a79db8109c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.644009] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5267f578-112e-86b6-3e07-fb4ce08852c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010377} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.644821] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90146780-5ff1-4f28-b5e7-201c2940aad4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.650751] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1268.650751] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52846b57-531b-ec4e-2a82-913ed9869020" [ 1268.650751] env[65385]: _type = "Task" [ 1268.650751] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.659688] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52846b57-531b-ec4e-2a82-913ed9869020, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.161986] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52846b57-531b-ec4e-2a82-913ed9869020, 'name': SearchDatastore_Task, 'duration_secs': 0.012485} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.162214] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1269.162457] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 6926f1bc-3c40-486f-b3e9-7e3731d57e4e/6926f1bc-3c40-486f-b3e9-7e3731d57e4e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1269.162726] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0df90326-050b-4e40-b6d5-15ee97412505 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.170026] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1269.170026] env[65385]: value = "task-4454717" [ 1269.170026] env[65385]: _type = "Task" [ 1269.170026] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.178404] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454717, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.483905] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.484140] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.684928] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454717, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.986762] env[65385]: DEBUG nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1270.183566] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454717, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.503598] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1270.503850] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1270.506586] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1270.506834] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1270.508701] env[65385]: INFO nova.compute.claims [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1270.681437] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454717, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.006852] env[65385]: DEBUG nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1271.182398] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454717, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.552737} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.182652] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 6926f1bc-3c40-486f-b3e9-7e3731d57e4e/6926f1bc-3c40-486f-b3e9-7e3731d57e4e.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1271.182865] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1271.183146] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1e24e66-7412-4fc5-8d1f-563de59a9c94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.191321] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1271.191321] env[65385]: value = "task-4454718" [ 1271.191321] env[65385]: _type = "Task" [ 1271.191321] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.202401] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454718, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.530034] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1271.602244] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d7c039-991e-48d4-859c-bf98e17ca887 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.611057] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97dda61b-a025-4244-aef2-9f62e48a5124 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.642511] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ec4973-3f81-49cc-816c-c319b46aeb86 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.650297] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f90d8c-e3fb-4230-8ce2-6bb6154acaba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.666080] env[65385]: DEBUG nova.compute.provider_tree [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1271.700981] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454718, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093528} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.701254] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1271.702019] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef5c2c5-76c7-4f9f-90e4-85fd144032bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.724093] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] 6926f1bc-3c40-486f-b3e9-7e3731d57e4e/6926f1bc-3c40-486f-b3e9-7e3731d57e4e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1271.724371] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cab5c1d-f352-459a-af2d-285b3e78e4e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.744096] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1271.744096] env[65385]: value = "task-4454719" [ 1271.744096] env[65385]: _type = "Task" [ 1271.744096] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.752390] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454719, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.169257] env[65385]: DEBUG nova.scheduler.client.report [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1272.254487] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454719, 'name': ReconfigVM_Task, 'duration_secs': 0.29119} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.254719] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Reconfigured VM instance instance-00000079 to attach disk [datastore1] 6926f1bc-3c40-486f-b3e9-7e3731d57e4e/6926f1bc-3c40-486f-b3e9-7e3731d57e4e.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1272.255394] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc76e041-2bc0-413e-9b6c-99e6e96318cd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.263270] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1272.263270] env[65385]: value = "task-4454720" [ 1272.263270] env[65385]: _type = "Task" [ 1272.263270] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.271813] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454720, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.673848] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.167s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1272.674211] env[65385]: DEBUG nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1272.677072] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.148s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1272.678559] env[65385]: INFO nova.compute.claims [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1272.773544] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454720, 'name': Rename_Task, 'duration_secs': 0.162361} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.773809] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1272.774075] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e84f682-928c-4c70-b88f-2f5bb6cc7bf5 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.781232] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1272.781232] env[65385]: value = "task-4454721" [ 1272.781232] env[65385]: _type = "Task" [ 1272.781232] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.789572] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454721, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.183534] env[65385]: DEBUG nova.compute.utils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1273.187256] env[65385]: DEBUG nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1273.187468] env[65385]: DEBUG nova.network.neutron [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1273.187791] env[65385]: WARNING neutronclient.v2_0.client [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1273.188098] env[65385]: WARNING neutronclient.v2_0.client [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1273.188674] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1273.189017] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1273.244835] env[65385]: DEBUG nova.policy [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2b450c500f84f2fa682d70698c398b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d45ff3446a84ef2a1945f7a66b747e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1273.292011] env[65385]: DEBUG oslo_vmware.api [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454721, 'name': PowerOnVM_Task, 'duration_secs': 0.471205} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.292307] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1273.292508] env[65385]: INFO nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Took 8.31 seconds to spawn the instance on the hypervisor. [ 1273.292677] env[65385]: DEBUG nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1273.293512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a3fa0c-b69d-49cb-be84-af49f0429396 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.577067] env[65385]: DEBUG nova.network.neutron [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Successfully created port: 5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1273.688439] env[65385]: DEBUG nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1273.809436] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f2e707-2abc-4df1-b414-759eff5434db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.814426] env[65385]: INFO nova.compute.manager [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Took 13.03 seconds to build instance. [ 1273.819314] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a87c01d-de5a-469b-a8c5-a9a6aa38aeb2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.851219] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a01ee44-eeca-4902-b88a-5526e41cab78 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.860180] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5eadd6f-1567-4d07-a1cb-44ba272dc694 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.875080] env[65385]: DEBUG nova.compute.provider_tree [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1274.064568] env[65385]: DEBUG nova.compute.manager [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Received event network-changed-59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1274.064773] env[65385]: DEBUG nova.compute.manager [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Refreshing instance network info cache due to event network-changed-59e7b718-2455-470d-85a3-3a0d0849ea44. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1274.064981] env[65385]: DEBUG oslo_concurrency.lockutils [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Acquiring lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.065182] env[65385]: DEBUG oslo_concurrency.lockutils [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Acquired lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1274.065386] env[65385]: DEBUG nova.network.neutron [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Refreshing network info cache for port 59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1274.195019] env[65385]: INFO nova.virt.block_device [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Booting with volume abe0b31f-295c-4222-ba80-83d0bae39141 at /dev/sda [ 1274.235907] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a1f9a47-f245-4d3a-b317-5735e4658237 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.246914] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6507351a-7b82-4cfe-983f-515a6ad25f12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.274965] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4a5a0df-c543-4d69-966a-9e57935d9be2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.284915] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8007cc48-b900-4d0c-865c-4da315a06906 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.313537] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8c72bc-ffea-477c-9e82-1d06274693a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.316369] env[65385]: DEBUG oslo_concurrency.lockutils [None req-37b7d61c-0b2f-4bce-bfd5-9875c166cf30 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.540s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1274.321270] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e99909-ec74-4fce-b706-54a6a2f809d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.334900] env[65385]: DEBUG nova.virt.block_device [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating existing volume attachment record: 595826c6-3abf-40bf-8816-bc89eddbe039 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1274.378109] env[65385]: DEBUG nova.scheduler.client.report [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1274.568152] env[65385]: WARNING neutronclient.v2_0.client [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.569665] env[65385]: WARNING openstack [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.570072] env[65385]: WARNING openstack [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.695559] env[65385]: WARNING openstack [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.695997] env[65385]: WARNING openstack [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.759597] env[65385]: WARNING neutronclient.v2_0.client [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.760371] env[65385]: WARNING openstack [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.760746] env[65385]: WARNING openstack [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.844831] env[65385]: DEBUG nova.network.neutron [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updated VIF entry in instance network info cache for port 59e7b718-2455-470d-85a3-3a0d0849ea44. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1274.845200] env[65385]: DEBUG nova.network.neutron [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating instance_info_cache with network_info: [{"id": "59e7b718-2455-470d-85a3-3a0d0849ea44", "address": "fa:16:3e:7e:65:d9", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e7b718-24", "ovs_interfaceid": "59e7b718-2455-470d-85a3-3a0d0849ea44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1274.882689] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1274.883219] env[65385]: DEBUG nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1275.347712] env[65385]: DEBUG oslo_concurrency.lockutils [req-6d455a1b-7406-41ae-8ec5-818ad629ad38 req-967943b2-74f9-4917-9859-78916126060b service nova] Releasing lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1275.388349] env[65385]: DEBUG nova.compute.utils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1275.389940] env[65385]: DEBUG nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1275.394026] env[65385]: DEBUG nova.network.neutron [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1275.394026] env[65385]: WARNING neutronclient.v2_0.client [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.394026] env[65385]: WARNING neutronclient.v2_0.client [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.394026] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1275.394026] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1275.422024] env[65385]: DEBUG nova.network.neutron [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Successfully updated port: 5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1275.454938] env[65385]: DEBUG nova.compute.manager [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Received event network-changed-5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1275.454938] env[65385]: DEBUG nova.compute.manager [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Refreshing instance network info cache due to event network-changed-5ce9de22-0e36-4aa4-86cc-4794bae3842c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1275.455094] env[65385]: DEBUG oslo_concurrency.lockutils [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] Acquiring lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.455597] env[65385]: DEBUG oslo_concurrency.lockutils [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] Acquired lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1275.455597] env[65385]: DEBUG nova.network.neutron [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Refreshing network info cache for port 5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1275.465414] env[65385]: DEBUG nova.policy [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '043f66592e52480c87c36a10799be03c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96214fc566ed4b11a92a6369a3419aea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1275.595199] env[65385]: DEBUG nova.compute.manager [req-0c648912-a79c-4afe-ab0b-4a725aa2b66f req-e0c40409-890c-4081-b350-8187fdde0a8b service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Received event network-vif-plugged-5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1275.595199] env[65385]: DEBUG oslo_concurrency.lockutils [req-0c648912-a79c-4afe-ab0b-4a725aa2b66f req-e0c40409-890c-4081-b350-8187fdde0a8b service nova] Acquiring lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1275.595389] env[65385]: DEBUG oslo_concurrency.lockutils [req-0c648912-a79c-4afe-ab0b-4a725aa2b66f req-e0c40409-890c-4081-b350-8187fdde0a8b service nova] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1275.596139] env[65385]: DEBUG oslo_concurrency.lockutils [req-0c648912-a79c-4afe-ab0b-4a725aa2b66f req-e0c40409-890c-4081-b350-8187fdde0a8b service nova] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1275.596139] env[65385]: DEBUG nova.compute.manager [req-0c648912-a79c-4afe-ab0b-4a725aa2b66f req-e0c40409-890c-4081-b350-8187fdde0a8b service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] No waiting events found dispatching network-vif-plugged-5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1275.596139] env[65385]: WARNING nova.compute.manager [req-0c648912-a79c-4afe-ab0b-4a725aa2b66f req-e0c40409-890c-4081-b350-8187fdde0a8b service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Received unexpected event network-vif-plugged-5ce9de22-0e36-4aa4-86cc-4794bae3842c for instance with vm_state building and task_state block_device_mapping. [ 1275.786334] env[65385]: DEBUG nova.network.neutron [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Successfully created port: 0674090d-6906-4239-990b-c62867c7dce0 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1275.904608] env[65385]: DEBUG nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1275.926221] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.960243] env[65385]: WARNING neutronclient.v2_0.client [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.961389] env[65385]: WARNING openstack [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1275.961992] env[65385]: WARNING openstack [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1275.998547] env[65385]: DEBUG nova.network.neutron [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1276.085753] env[65385]: DEBUG nova.network.neutron [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1276.441008] env[65385]: DEBUG nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1276.441595] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1276.441859] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1276.442038] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1276.442272] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1276.442367] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1276.442611] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1276.442797] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1276.442975] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1276.443159] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1276.443322] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1276.443507] env[65385]: DEBUG nova.virt.hardware [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1276.444494] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26119fe6-1a7d-463d-9e01-207039a915b1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.454873] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35c3934-74f7-421d-9763-c5af5519f5f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.588922] env[65385]: DEBUG oslo_concurrency.lockutils [req-078d1285-766b-435c-b256-2ddc12313f35 req-e04003e1-3f96-4e0f-b3ef-96356587633d service nova] Releasing lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1276.589447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1276.589641] env[65385]: DEBUG nova.network.neutron [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1276.914734] env[65385]: DEBUG nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1276.943902] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1276.944185] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1276.944408] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1276.944609] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1276.944754] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1276.944895] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1276.945118] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1276.945360] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1276.945554] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1276.945721] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1276.945892] env[65385]: DEBUG nova.virt.hardware [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1276.946785] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166d24f9-b310-4cee-973f-70235360c852 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.955925] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5ba854-50ff-4dc0-8d1f-8569207990ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.092695] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.093186] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.128836] env[65385]: DEBUG nova.network.neutron [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1277.148249] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.148699] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.205061] env[65385]: DEBUG nova.compute.manager [req-47ca4c46-deff-4e91-9404-307234c523dd req-a684b5da-35bc-4249-8719-04964f3b197c service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Received event network-vif-plugged-0674090d-6906-4239-990b-c62867c7dce0 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1277.205248] env[65385]: DEBUG oslo_concurrency.lockutils [req-47ca4c46-deff-4e91-9404-307234c523dd req-a684b5da-35bc-4249-8719-04964f3b197c service nova] Acquiring lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1277.205456] env[65385]: DEBUG oslo_concurrency.lockutils [req-47ca4c46-deff-4e91-9404-307234c523dd req-a684b5da-35bc-4249-8719-04964f3b197c service nova] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1277.205620] env[65385]: DEBUG oslo_concurrency.lockutils [req-47ca4c46-deff-4e91-9404-307234c523dd req-a684b5da-35bc-4249-8719-04964f3b197c service nova] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1277.206255] env[65385]: DEBUG nova.compute.manager [req-47ca4c46-deff-4e91-9404-307234c523dd req-a684b5da-35bc-4249-8719-04964f3b197c service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] No waiting events found dispatching network-vif-plugged-0674090d-6906-4239-990b-c62867c7dce0 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1277.206255] env[65385]: WARNING nova.compute.manager [req-47ca4c46-deff-4e91-9404-307234c523dd req-a684b5da-35bc-4249-8719-04964f3b197c service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Received unexpected event network-vif-plugged-0674090d-6906-4239-990b-c62867c7dce0 for instance with vm_state building and task_state spawning. [ 1277.212852] env[65385]: WARNING neutronclient.v2_0.client [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1277.213521] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.213882] env[65385]: WARNING openstack [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.291034] env[65385]: DEBUG nova.network.neutron [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Successfully updated port: 0674090d-6906-4239-990b-c62867c7dce0 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1277.296163] env[65385]: DEBUG nova.network.neutron [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [{"id": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "address": "fa:16:3e:57:7e:6a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ce9de22-0e", "ovs_interfaceid": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1277.791942] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1277.792323] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1277.792483] env[65385]: DEBUG nova.network.neutron [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1277.798861] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1277.799196] env[65385]: DEBUG nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Instance network_info: |[{"id": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "address": "fa:16:3e:57:7e:6a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ce9de22-0e", "ovs_interfaceid": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1277.799638] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:7e:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ce9de22-0e36-4aa4-86cc-4794bae3842c', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1277.808942] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1277.809812] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1277.810170] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f795ffc9-177c-4478-b745-fff3aeff231a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.831168] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1277.831168] env[65385]: value = "task-4454722" [ 1277.831168] env[65385]: _type = "Task" [ 1277.831168] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.839998] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454722, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.295859] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.296350] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.339226] env[65385]: DEBUG nova.network.neutron [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1278.344859] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454722, 'name': CreateVM_Task, 'duration_secs': 0.332472} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.345032] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1278.345806] env[65385]: WARNING neutronclient.v2_0.client [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.345851] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871211', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'name': 'volume-abe0b31f-295c-4222-ba80-83d0bae39141', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2c15094a-1c22-4e06-baf1-1be4a6fb432a', 'attached_at': '', 'detached_at': '', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'serial': 'abe0b31f-295c-4222-ba80-83d0bae39141'}, 'device_type': None, 'guest_format': None, 'attachment_id': '595826c6-3abf-40bf-8816-bc89eddbe039', 'mount_device': '/dev/sda', 'disk_bus': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=65385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1278.346044] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Root volume attach. Driver type: vmdk {{(pid=65385) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1278.346823] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea6eabf-e6d1-4f83-bdc3-33edb1b6262c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.355830] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ca7501-8f0c-455e-92f9-504c9e53471e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.361624] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.362054] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.370870] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-902bbba3-574b-4ddb-94f2-f5c661626faf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.377951] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8ad83804-cc88-4d5d-9b2c-87183d33f621 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.386248] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1278.386248] env[65385]: value = "task-4454723" [ 1278.386248] env[65385]: _type = "Task" [ 1278.386248] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.398085] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454723, 'name': RelocateVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.412984] env[65385]: WARNING neutronclient.v2_0.client [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.413688] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.414412] env[65385]: WARNING openstack [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.501369] env[65385]: DEBUG nova.network.neutron [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Updating instance_info_cache with network_info: [{"id": "0674090d-6906-4239-990b-c62867c7dce0", "address": "fa:16:3e:0f:bd:24", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0674090d-69", "ovs_interfaceid": "0674090d-6906-4239-990b-c62867c7dce0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1278.896321] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454723, 'name': RelocateVM_Task, 'duration_secs': 0.380157} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.896691] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1278.896732] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871211', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'name': 'volume-abe0b31f-295c-4222-ba80-83d0bae39141', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2c15094a-1c22-4e06-baf1-1be4a6fb432a', 'attached_at': '', 'detached_at': '', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'serial': 'abe0b31f-295c-4222-ba80-83d0bae39141'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1278.897511] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980d6324-5c0b-4f5c-98b8-ddc9776db87d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.912856] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfe8605-ae3a-41af-8229-e45f39a17b1e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.935061] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] volume-abe0b31f-295c-4222-ba80-83d0bae39141/volume-abe0b31f-295c-4222-ba80-83d0bae39141.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1278.935413] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-812ce98d-5bf1-4fce-aaa5-08dc284c7da9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.958582] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1278.958582] env[65385]: value = "task-4454724" [ 1278.958582] env[65385]: _type = "Task" [ 1278.958582] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.966956] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454724, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.004969] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1279.005422] env[65385]: DEBUG nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Instance network_info: |[{"id": "0674090d-6906-4239-990b-c62867c7dce0", "address": "fa:16:3e:0f:bd:24", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0674090d-69", "ovs_interfaceid": "0674090d-6906-4239-990b-c62867c7dce0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1279.005932] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:bd:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe99da4f-5630-4afd-918b-b327193d8489', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0674090d-6906-4239-990b-c62867c7dce0', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1279.013703] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1279.013918] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1279.014180] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-411c00a5-c4e7-43fc-9a69-74a1c227c00d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.035594] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1279.035594] env[65385]: value = "task-4454725" [ 1279.035594] env[65385]: _type = "Task" [ 1279.035594] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.043813] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454725, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.244363] env[65385]: DEBUG nova.compute.manager [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Received event network-changed-0674090d-6906-4239-990b-c62867c7dce0 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1279.244557] env[65385]: DEBUG nova.compute.manager [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Refreshing instance network info cache due to event network-changed-0674090d-6906-4239-990b-c62867c7dce0. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1279.244861] env[65385]: DEBUG oslo_concurrency.lockutils [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Acquiring lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.245045] env[65385]: DEBUG oslo_concurrency.lockutils [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Acquired lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1279.245228] env[65385]: DEBUG nova.network.neutron [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Refreshing network info cache for port 0674090d-6906-4239-990b-c62867c7dce0 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1279.468940] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454724, 'name': ReconfigVM_Task, 'duration_secs': 0.296867} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.469269] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfigured VM instance instance-0000007a to attach disk [datastore1] volume-abe0b31f-295c-4222-ba80-83d0bae39141/volume-abe0b31f-295c-4222-ba80-83d0bae39141.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1279.474082] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a362358-3835-48c0-91c3-981a41d425e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.490075] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1279.490075] env[65385]: value = "task-4454726" [ 1279.490075] env[65385]: _type = "Task" [ 1279.490075] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.498673] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454726, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.545745] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454725, 'name': CreateVM_Task, 'duration_secs': 0.307539} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.545926] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1279.546498] env[65385]: WARNING neutronclient.v2_0.client [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1279.546921] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.547119] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1279.547442] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1279.547702] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a6ccf1c-7682-4bbe-96f2-ba5f88ab252d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.553020] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1279.553020] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52273374-9d4d-c098-c5db-4d1913b5e8c1" [ 1279.553020] env[65385]: _type = "Task" [ 1279.553020] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.561628] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52273374-9d4d-c098-c5db-4d1913b5e8c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.748751] env[65385]: WARNING neutronclient.v2_0.client [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1279.749605] env[65385]: WARNING openstack [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.750128] env[65385]: WARNING openstack [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1279.841594] env[65385]: WARNING openstack [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.842008] env[65385]: WARNING openstack [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1279.888710] env[65385]: WARNING neutronclient.v2_0.client [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1279.889370] env[65385]: WARNING openstack [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.889731] env[65385]: WARNING openstack [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1279.972068] env[65385]: DEBUG nova.network.neutron [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Updated VIF entry in instance network info cache for port 0674090d-6906-4239-990b-c62867c7dce0. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1279.972444] env[65385]: DEBUG nova.network.neutron [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Updating instance_info_cache with network_info: [{"id": "0674090d-6906-4239-990b-c62867c7dce0", "address": "fa:16:3e:0f:bd:24", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0674090d-69", "ovs_interfaceid": "0674090d-6906-4239-990b-c62867c7dce0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1280.001368] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454726, 'name': ReconfigVM_Task, 'duration_secs': 0.127244} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.001721] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871211', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'name': 'volume-abe0b31f-295c-4222-ba80-83d0bae39141', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2c15094a-1c22-4e06-baf1-1be4a6fb432a', 'attached_at': '', 'detached_at': '', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'serial': 'abe0b31f-295c-4222-ba80-83d0bae39141'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1280.002426] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0e95e8a-f722-45cb-9a82-31c7e07fe26d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.009350] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1280.009350] env[65385]: value = "task-4454727" [ 1280.009350] env[65385]: _type = "Task" [ 1280.009350] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.017822] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454727, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.064023] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52273374-9d4d-c098-c5db-4d1913b5e8c1, 'name': SearchDatastore_Task, 'duration_secs': 0.011039} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.064388] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1280.064632] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1280.064895] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.065015] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1280.065198] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1280.065480] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32fcbf40-5e7a-48be-802b-147858885030 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.074600] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1280.074769] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1280.075505] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dacbe1d-0af0-4018-8e0d-36c2efb192ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.081472] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1280.081472] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52decb21-d067-3ef6-e364-7c9d7ec0f716" [ 1280.081472] env[65385]: _type = "Task" [ 1280.081472] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.089696] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52decb21-d067-3ef6-e364-7c9d7ec0f716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.475942] env[65385]: DEBUG oslo_concurrency.lockutils [req-f9d30c3a-6046-4a6b-aa21-44671deb3d18 req-41448b99-71d2-4dea-9f8a-ec95c55e7663 service nova] Releasing lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1280.520920] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454727, 'name': Rename_Task, 'duration_secs': 0.214921} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.521214] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1280.521457] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c06a43ce-4e23-4b3a-9a6e-20d1c1361ec6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.527536] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1280.527536] env[65385]: value = "task-4454728" [ 1280.527536] env[65385]: _type = "Task" [ 1280.527536] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.535585] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.592297] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52decb21-d067-3ef6-e364-7c9d7ec0f716, 'name': SearchDatastore_Task, 'duration_secs': 0.009442} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.593129] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8afa74cf-0a92-4f9f-9f64-4de6c0af9a77 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.599686] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1280.599686] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c134d-b482-97da-c964-886837e9b577" [ 1280.599686] env[65385]: _type = "Task" [ 1280.599686] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.608658] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c134d-b482-97da-c964-886837e9b577, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.037632] env[65385]: DEBUG oslo_vmware.api [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454728, 'name': PowerOnVM_Task, 'duration_secs': 0.498442} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.038029] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1281.038214] env[65385]: INFO nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Took 4.60 seconds to spawn the instance on the hypervisor. [ 1281.038516] env[65385]: DEBUG nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1281.039460] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30365f0b-b258-4d53-a588-94ddcde0b017 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.111590] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]520c134d-b482-97da-c964-886837e9b577, 'name': SearchDatastore_Task, 'duration_secs': 0.014995} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.111794] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1281.112058] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/5d2820b9-db76-4572-b0ab-3c4a3ced0fdb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1281.112350] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-351aa296-379c-41cc-83a0-0f6473513785 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.120186] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1281.120186] env[65385]: value = "task-4454729" [ 1281.120186] env[65385]: _type = "Task" [ 1281.120186] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.129702] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454729, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.558955] env[65385]: INFO nova.compute.manager [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Took 11.07 seconds to build instance. [ 1281.630670] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454729, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47424} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.630923] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/5d2820b9-db76-4572-b0ab-3c4a3ced0fdb.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1281.631144] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1281.631393] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d60b4a2-23a0-476d-bb0b-abdac3a2d2d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.638526] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1281.638526] env[65385]: value = "task-4454730" [ 1281.638526] env[65385]: _type = "Task" [ 1281.638526] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.647170] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454730, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.061049] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4d7700fa-afc4-44a0-a3f3-c3eccb49df95 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.577s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1282.148762] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454730, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074653} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.148992] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1282.149896] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a2ac45-9ac2-4e80-b523-4cf65796b75d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.173295] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/5d2820b9-db76-4572-b0ab-3c4a3ced0fdb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1282.173644] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1d14468-c862-4a3b-8877-c353844f20c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.194366] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1282.194366] env[65385]: value = "task-4454731" [ 1282.194366] env[65385]: _type = "Task" [ 1282.194366] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.203321] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454731, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.714776] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454731, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.745466] env[65385]: DEBUG nova.compute.manager [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Received event network-changed-c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1282.745816] env[65385]: DEBUG nova.compute.manager [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Refreshing instance network info cache due to event network-changed-c11bab1f-7eb9-434c-8436-c2319951b01f. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1282.745939] env[65385]: DEBUG oslo_concurrency.lockutils [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Acquiring lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.746080] env[65385]: DEBUG oslo_concurrency.lockutils [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Acquired lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1282.746236] env[65385]: DEBUG nova.network.neutron [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Refreshing network info cache for port c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1283.206795] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454731, 'name': ReconfigVM_Task, 'duration_secs': 0.886661} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.207247] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Reconfigured VM instance instance-0000007b to attach disk [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/5d2820b9-db76-4572-b0ab-3c4a3ced0fdb.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1283.207722] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2839af71-04e1-4997-b04d-73e9cd0fe820 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.214623] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1283.214623] env[65385]: value = "task-4454732" [ 1283.214623] env[65385]: _type = "Task" [ 1283.214623] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.223664] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454732, 'name': Rename_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.248848] env[65385]: WARNING neutronclient.v2_0.client [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.249554] env[65385]: WARNING openstack [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.249821] env[65385]: WARNING openstack [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.371490] env[65385]: WARNING openstack [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.371490] env[65385]: WARNING openstack [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.434502] env[65385]: WARNING neutronclient.v2_0.client [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.435174] env[65385]: WARNING openstack [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.435529] env[65385]: WARNING openstack [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.515770] env[65385]: DEBUG nova.network.neutron [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updated VIF entry in instance network info cache for port c11bab1f-7eb9-434c-8436-c2319951b01f. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1283.516154] env[65385]: DEBUG nova.network.neutron [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updating instance_info_cache with network_info: [{"id": "c11bab1f-7eb9-434c-8436-c2319951b01f", "address": "fa:16:3e:ea:99:f4", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc11bab1f-7e", "ovs_interfaceid": "c11bab1f-7eb9-434c-8436-c2319951b01f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1283.725855] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454732, 'name': Rename_Task, 'duration_secs': 0.14692} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.725855] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1283.725855] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afedeffe-5db0-4a78-afca-466d29a3c296 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.733633] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1283.733633] env[65385]: value = "task-4454733" [ 1283.733633] env[65385]: _type = "Task" [ 1283.733633] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.742798] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.018896] env[65385]: DEBUG oslo_concurrency.lockutils [req-aedf30ea-7055-4ff0-a1e2-94bbaa10f78c req-75db7383-528a-46da-be80-f02323dad73a service nova] Releasing lock "refresh_cache-b4bca2c8-fc8f-471a-899b-aac648fee9fe" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.244836] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454733, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.266756] env[65385]: DEBUG nova.compute.manager [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Stashing vm_state: active {{(pid=65385) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1284.745704] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454733, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.768223] env[65385]: DEBUG nova.compute.manager [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Received event network-changed-5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1284.768431] env[65385]: DEBUG nova.compute.manager [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Refreshing instance network info cache due to event network-changed-5ce9de22-0e36-4aa4-86cc-4794bae3842c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1284.768638] env[65385]: DEBUG oslo_concurrency.lockutils [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Acquiring lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.768773] env[65385]: DEBUG oslo_concurrency.lockutils [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Acquired lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1284.768926] env[65385]: DEBUG nova.network.neutron [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Refreshing network info cache for port 5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1284.787330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1284.787759] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1285.247418] env[65385]: DEBUG oslo_vmware.api [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454733, 'name': PowerOnVM_Task, 'duration_secs': 1.423571} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.247838] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1285.247838] env[65385]: INFO nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Took 8.33 seconds to spawn the instance on the hypervisor. [ 1285.248032] env[65385]: DEBUG nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1285.248812] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7170390e-e136-413d-bffa-9a51ecc16550 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.275051] env[65385]: WARNING neutronclient.v2_0.client [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1285.275356] env[65385]: WARNING openstack [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1285.275723] env[65385]: WARNING openstack [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1285.292389] env[65385]: INFO nova.compute.claims [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1285.412604] env[65385]: WARNING openstack [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1285.413021] env[65385]: WARNING openstack [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1285.487514] env[65385]: WARNING neutronclient.v2_0.client [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1285.488181] env[65385]: WARNING openstack [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1285.488521] env[65385]: WARNING openstack [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1285.572816] env[65385]: DEBUG nova.network.neutron [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updated VIF entry in instance network info cache for port 5ce9de22-0e36-4aa4-86cc-4794bae3842c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1285.573203] env[65385]: DEBUG nova.network.neutron [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [{"id": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "address": "fa:16:3e:57:7e:6a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ce9de22-0e", "ovs_interfaceid": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1285.765797] env[65385]: INFO nova.compute.manager [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Took 14.25 seconds to build instance. [ 1285.799626] env[65385]: INFO nova.compute.resource_tracker [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating resource usage from migration 7448bd75-40bf-489b-b165-5ae8206b9677 [ 1285.901096] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9609d5-7324-4cdd-860b-2a683cf6e8c4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.909401] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a5172f-3ea0-435a-bb0b-e2815152bd13 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.940780] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d8c29f-c206-4e6b-aaf7-3b0a76b3756a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.949255] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511a00fc-c32b-495b-bb77-7b4bce297777 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.965917] env[65385]: DEBUG nova.compute.provider_tree [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1286.076384] env[65385]: DEBUG oslo_concurrency.lockutils [req-8f62df97-735c-42d8-a302-593dc5209d68 req-b72b70c5-b69d-4d42-b8fc-32939c44dc03 service nova] Releasing lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1286.268412] env[65385]: DEBUG oslo_concurrency.lockutils [None req-cf94bc23-beec-4955-b59d-c48e16f42fb9 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.764s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1286.469877] env[65385]: DEBUG nova.scheduler.client.report [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1286.975684] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.188s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1286.975933] env[65385]: INFO nova.compute.manager [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Migrating [ 1287.060451] env[65385]: INFO nova.compute.manager [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Rescuing [ 1287.060719] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.060865] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1287.061035] env[65385]: DEBUG nova.network.neutron [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1287.491382] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.491783] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1287.491783] env[65385]: DEBUG nova.network.neutron [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1287.564207] env[65385]: WARNING neutronclient.v2_0.client [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1287.564986] env[65385]: WARNING openstack [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1287.565358] env[65385]: WARNING openstack [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1287.664717] env[65385]: WARNING openstack [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1287.665232] env[65385]: WARNING openstack [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1287.714242] env[65385]: WARNING neutronclient.v2_0.client [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1287.715160] env[65385]: WARNING openstack [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1287.715308] env[65385]: WARNING openstack [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1287.793969] env[65385]: DEBUG nova.network.neutron [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Updating instance_info_cache with network_info: [{"id": "0674090d-6906-4239-990b-c62867c7dce0", "address": "fa:16:3e:0f:bd:24", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0674090d-69", "ovs_interfaceid": "0674090d-6906-4239-990b-c62867c7dce0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1287.995192] env[65385]: WARNING neutronclient.v2_0.client [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1287.995887] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1287.996307] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.114373] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.114877] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.187358] env[65385]: WARNING neutronclient.v2_0.client [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1288.188013] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.188397] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.280241] env[65385]: DEBUG nova.network.neutron [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [{"id": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "address": "fa:16:3e:57:7e:6a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ce9de22-0e", "ovs_interfaceid": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1288.296448] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1288.783326] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1289.839512] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1289.839860] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ffa98f2-7fd6-4756-b774-f55d8e31a6a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.848390] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1289.848390] env[65385]: value = "task-4454734" [ 1289.848390] env[65385]: _type = "Task" [ 1289.848390] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.857180] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.299998] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bce57ea-5ef2-4bcd-b482-615ce72962fa {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.320167] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance '2c15094a-1c22-4e06-baf1-1be4a6fb432a' progress to 0 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1290.359217] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454734, 'name': PowerOffVM_Task, 'duration_secs': 0.206969} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.359489] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1290.360467] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7f2ab3-6591-4859-a070-4a68f86a6dcb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.379584] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f704c1-11b1-4a13-84b0-28674309b915 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.413701] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1290.413993] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e553d13-8891-47ae-94cf-8d1ee2d400ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.421467] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1290.421467] env[65385]: value = "task-4454735" [ 1290.421467] env[65385]: _type = "Task" [ 1290.421467] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.429846] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.826582] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1290.826902] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7310626a-9ed3-4cb1-8ca2-204e77425275 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.834790] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1290.834790] env[65385]: value = "task-4454736" [ 1290.834790] env[65385]: _type = "Task" [ 1290.834790] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.843375] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454736, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.936661] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1290.936661] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1290.936936] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1290.937091] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1290.937278] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1290.937597] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5ff8250-e0ba-4fc4-8c24-090b95757166 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.950015] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1290.950258] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1290.950990] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5685b609-45d8-415a-a176-a3932798b7f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.958556] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1290.958556] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5242bc9b-f153-5e6b-55f5-817cfbd55171" [ 1290.958556] env[65385]: _type = "Task" [ 1290.958556] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.966438] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5242bc9b-f153-5e6b-55f5-817cfbd55171, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.345773] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454736, 'name': PowerOffVM_Task, 'duration_secs': 0.171622} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.346290] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1291.346614] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance '2c15094a-1c22-4e06-baf1-1be4a6fb432a' progress to 17 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1291.468408] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5242bc9b-f153-5e6b-55f5-817cfbd55171, 'name': SearchDatastore_Task, 'duration_secs': 0.010619} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.469472] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaa8db65-f4bc-4088-b31d-1df301959dd2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.475822] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1291.475822] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae1db0-8901-5ff2-d678-78f501291be0" [ 1291.475822] env[65385]: _type = "Task" [ 1291.475822] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.484818] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae1db0-8901-5ff2-d678-78f501291be0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.853054] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1291.853478] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1291.853478] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1291.853623] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1291.853721] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1291.853863] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1291.854081] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1291.854237] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1291.854395] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1291.854578] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1291.854761] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1291.859698] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6ebaa49-6fca-4353-b5bc-3b73c13ef8bd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.875795] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1291.875795] env[65385]: value = "task-4454737" [ 1291.875795] env[65385]: _type = "Task" [ 1291.875795] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.885057] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454737, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.987409] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52ae1db0-8901-5ff2-d678-78f501291be0, 'name': SearchDatastore_Task, 'duration_secs': 0.010538} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.987665] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1291.987918] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. {{(pid=65385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1291.988217] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef62de45-fcf5-418d-a454-f70196a1617c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.995672] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1291.995672] env[65385]: value = "task-4454738" [ 1291.995672] env[65385]: _type = "Task" [ 1291.995672] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.005234] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.387041] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454737, 'name': ReconfigVM_Task, 'duration_secs': 0.172103} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.387445] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance '2c15094a-1c22-4e06-baf1-1be4a6fb432a' progress to 33 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1292.506256] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477291} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.506743] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. [ 1292.507589] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c944666-5845-4c8f-a483-6c221c6fe17c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.532807] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1292.533180] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a790086-d91a-48df-9200-76c662dafb3f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.551352] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1292.551352] env[65385]: value = "task-4454739" [ 1292.551352] env[65385]: _type = "Task" [ 1292.551352] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.561260] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.894037] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1292.894592] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1292.894592] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1292.894685] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1292.894800] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1292.894937] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1292.895153] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1292.895309] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1292.895504] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1292.896078] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1292.896078] env[65385]: DEBUG nova.virt.hardware [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1292.900978] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfiguring VM instance instance-0000007a to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1292.901293] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fedf327a-ad56-4d50-924d-086bf71be6f7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.921909] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1292.921909] env[65385]: value = "task-4454740" [ 1292.921909] env[65385]: _type = "Task" [ 1292.921909] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.930507] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454740, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.061692] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454739, 'name': ReconfigVM_Task, 'duration_secs': 0.267044} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.061994] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Reconfigured VM instance instance-0000007b to attach disk [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1293.062883] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bcff90-f44d-489e-9e4d-e5abf25974a4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.089049] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e02ba90-3e01-4c53-9761-b884d7fa27d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.105017] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1293.105017] env[65385]: value = "task-4454741" [ 1293.105017] env[65385]: _type = "Task" [ 1293.105017] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.114519] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454741, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.432301] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454740, 'name': ReconfigVM_Task, 'duration_secs': 0.170976} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.432609] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfigured VM instance instance-0000007a to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1293.433431] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200e373b-30cb-4041-bfae-eb1dc00d9c60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.455498] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] volume-abe0b31f-295c-4222-ba80-83d0bae39141/volume-abe0b31f-295c-4222-ba80-83d0bae39141.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1293.455749] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-991f5d20-6846-4d2d-a122-fe8205cff6c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.474533] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1293.474533] env[65385]: value = "task-4454742" [ 1293.474533] env[65385]: _type = "Task" [ 1293.474533] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.483297] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.615022] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454741, 'name': ReconfigVM_Task, 'duration_secs': 0.163597} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.615316] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1293.615588] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88909cbf-1033-43e6-87a4-7f5383214170 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.623475] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1293.623475] env[65385]: value = "task-4454743" [ 1293.623475] env[65385]: _type = "Task" [ 1293.623475] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.632055] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454743, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.985024] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454742, 'name': ReconfigVM_Task, 'duration_secs': 0.291865} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.985413] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfigured VM instance instance-0000007a to attach disk [datastore1] volume-abe0b31f-295c-4222-ba80-83d0bae39141/volume-abe0b31f-295c-4222-ba80-83d0bae39141.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1293.985594] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance '2c15094a-1c22-4e06-baf1-1be4a6fb432a' progress to 50 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1294.135104] env[65385]: DEBUG oslo_vmware.api [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454743, 'name': PowerOnVM_Task, 'duration_secs': 0.387023} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.135333] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1294.138063] env[65385]: DEBUG nova.compute.manager [None req-4e3adae2-3028-49ef-a36d-12c0f950f883 tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1294.138813] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d99a7b-f8f9-4cfb-89c4-75f61780037d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.492735] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219eabb8-e06e-45a9-b16b-44c65a634759 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.513356] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca65afa-098d-407a-8e05-bd8605d8bcc8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.532599] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance '2c15094a-1c22-4e06-baf1-1be4a6fb432a' progress to 67 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1295.042951] env[65385]: INFO nova.compute.manager [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Unrescuing [ 1295.043364] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1295.043472] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquired lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1295.043593] env[65385]: DEBUG nova.network.neutron [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1295.546693] env[65385]: WARNING neutronclient.v2_0.client [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1295.547413] env[65385]: WARNING openstack [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1295.547777] env[65385]: WARNING openstack [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1295.648593] env[65385]: WARNING openstack [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1295.649102] env[65385]: WARNING openstack [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1295.704376] env[65385]: WARNING neutronclient.v2_0.client [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1295.705093] env[65385]: WARNING openstack [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1295.705448] env[65385]: WARNING openstack [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1295.783748] env[65385]: DEBUG nova.network.neutron [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Updating instance_info_cache with network_info: [{"id": "0674090d-6906-4239-990b-c62867c7dce0", "address": "fa:16:3e:0f:bd:24", "network": {"id": "316139d1-ff31-4476-b823-3e0764de7675", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1692109322-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "96214fc566ed4b11a92a6369a3419aea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0674090d-69", "ovs_interfaceid": "0674090d-6906-4239-990b-c62867c7dce0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1296.160105] env[65385]: WARNING neutronclient.v2_0.client [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1296.160650] env[65385]: WARNING neutronclient.v2_0.client [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1296.196139] env[65385]: DEBUG nova.network.neutron [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Port 5ce9de22-0e36-4aa4-86cc-4794bae3842c binding to destination host cpu-1 is already ACTIVE {{(pid=65385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1296.286904] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Releasing lock "refresh_cache-5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1296.287568] env[65385]: DEBUG nova.objects.instance [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lazy-loading 'flavor' on Instance uuid 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1296.793206] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714c6d4a-309c-4bd2-a4ff-91aeed78d5ce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.817303] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1296.817643] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6545d740-4340-4053-b73c-ddaf2b89fef0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.825444] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1296.825444] env[65385]: value = "task-4454744" [ 1296.825444] env[65385]: _type = "Task" [ 1296.825444] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.834293] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.217583] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.217961] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.217961] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1297.335367] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454744, 'name': PowerOffVM_Task, 'duration_secs': 0.191874} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.335624] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1297.340827] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Reconfiguring VM instance instance-0000007b to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1297.341117] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a472aa0f-fce4-48a6-a8c0-44156f3bb00c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.360466] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1297.360466] env[65385]: value = "task-4454745" [ 1297.360466] env[65385]: _type = "Task" [ 1297.360466] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.369155] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454745, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.870549] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454745, 'name': ReconfigVM_Task, 'duration_secs': 0.205546} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.870756] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Reconfigured VM instance instance-0000007b to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1297.870937] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1297.871208] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1175089-4c04-4904-bc82-e40bf3aaafce {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.877845] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1297.877845] env[65385]: value = "task-4454746" [ 1297.877845] env[65385]: _type = "Task" [ 1297.877845] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.889237] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454746, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.223824] env[65385]: WARNING neutronclient.v2_0.client [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1298.259280] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1298.259516] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1298.259736] env[65385]: DEBUG nova.network.neutron [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1298.390056] env[65385]: DEBUG oslo_vmware.api [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454746, 'name': PowerOnVM_Task, 'duration_secs': 0.394276} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.390345] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1298.390568] env[65385]: DEBUG nova.compute.manager [None req-e9d15622-eeb4-4919-affe-aa84b982e7cf tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1298.391373] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e2bd37-ac7b-4eee-83a7-aa3f726ef5b9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.762523] env[65385]: WARNING neutronclient.v2_0.client [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1298.763879] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1298.763879] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1298.843777] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1298.883957] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1298.884373] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1298.954820] env[65385]: WARNING neutronclient.v2_0.client [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1298.955488] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1298.956014] env[65385]: WARNING openstack [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1299.035454] env[65385]: DEBUG nova.network.neutron [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [{"id": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "address": "fa:16:3e:57:7e:6a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ce9de22-0e", "ovs_interfaceid": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1299.538732] env[65385]: DEBUG oslo_concurrency.lockutils [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1299.660581] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1299.660819] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1299.661077] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1299.661314] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1299.661480] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1299.663893] env[65385]: INFO nova.compute.manager [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Terminating instance [ 1300.048929] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d04a495-18a0-4cc2-b318-c5bb9e0ceb12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.056980] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fb69ed-b531-417f-b075-cb3a6956e4ca {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.167632] env[65385]: DEBUG nova.compute.manager [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1300.167860] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1300.168788] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c5377d-7300-4be9-b49f-5f1a21d4d41b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.177331] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1300.177584] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e011896-af95-4512-8375-06c7047659c1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.184135] env[65385]: DEBUG oslo_vmware.api [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1300.184135] env[65385]: value = "task-4454747" [ 1300.184135] env[65385]: _type = "Task" [ 1300.184135] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.194031] env[65385]: DEBUG oslo_vmware.api [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454747, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.694582] env[65385]: DEBUG oslo_vmware.api [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454747, 'name': PowerOffVM_Task, 'duration_secs': 0.205478} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.694981] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1300.695032] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1300.695284] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dc22d82-d4d0-4286-8314-de15793a9054 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.764492] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1300.764773] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1300.764954] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Deleting the datastore file [datastore1] 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1300.765252] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38e11c37-0fe0-4e00-9293-81826f80a738 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.772925] env[65385]: DEBUG oslo_vmware.api [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1300.772925] env[65385]: value = "task-4454749" [ 1300.772925] env[65385]: _type = "Task" [ 1300.772925] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.782282] env[65385]: DEBUG oslo_vmware.api [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.159547] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4501c838-a537-4c24-90e5-2b362e84f206 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.181761] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19907625-faa4-46a9-a2ff-e63d537ae868 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.189741] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance '2c15094a-1c22-4e06-baf1-1be4a6fb432a' progress to 83 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1301.283320] env[65385]: DEBUG oslo_vmware.api [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134335} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.283550] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1301.283731] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1301.283899] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1301.284081] env[65385]: INFO nova.compute.manager [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1301.284319] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1301.284509] env[65385]: DEBUG nova.compute.manager [-] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1301.284635] env[65385]: DEBUG nova.network.neutron [-] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1301.284889] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1301.285420] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1301.285718] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1301.355712] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1301.617294] env[65385]: DEBUG nova.compute.manager [req-778e2e0a-d692-48d0-ac49-2357cabb4e5c req-fa72632f-1b9d-4b2b-9ac2-434321e53ca2 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Received event network-vif-deleted-0674090d-6906-4239-990b-c62867c7dce0 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1301.617514] env[65385]: INFO nova.compute.manager [req-778e2e0a-d692-48d0-ac49-2357cabb4e5c req-fa72632f-1b9d-4b2b-9ac2-434321e53ca2 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Neutron deleted interface 0674090d-6906-4239-990b-c62867c7dce0; detaching it from the instance and deleting it from the info cache [ 1301.617686] env[65385]: DEBUG nova.network.neutron [req-778e2e0a-d692-48d0-ac49-2357cabb4e5c req-fa72632f-1b9d-4b2b-9ac2-434321e53ca2 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1301.695908] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1301.696326] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-046d294f-ea10-451d-aad6-76bcc07e1975 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.706535] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1301.706535] env[65385]: value = "task-4454750" [ 1301.706535] env[65385]: _type = "Task" [ 1301.706535] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.715163] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454750, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.838128] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1301.843776] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1301.844037] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.098270] env[65385]: DEBUG nova.network.neutron [-] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1302.121147] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a617e01d-b613-4283-a402-110531572fbb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.131799] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb66150-df11-43d8-937c-cb7c4d113c35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.163160] env[65385]: DEBUG nova.compute.manager [req-778e2e0a-d692-48d0-ac49-2357cabb4e5c req-fa72632f-1b9d-4b2b-9ac2-434321e53ca2 service nova] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Detach interface failed, port_id=0674090d-6906-4239-990b-c62867c7dce0, reason: Instance 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1302.216595] env[65385]: DEBUG oslo_vmware.api [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454750, 'name': PowerOnVM_Task, 'duration_secs': 0.384289} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.216846] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1302.217031] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-2394d869-1d9f-4bf1-89e4-fb718abff918 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance '2c15094a-1c22-4e06-baf1-1be4a6fb432a' progress to 100 {{(pid=65385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1302.347112] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1302.347338] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1302.347509] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1302.347687] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1302.348577] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51b6a9b-52c8-4d6a-9b24-41b0b3d10b84 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.357316] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50750b65-22f5-40e4-9c0e-44737187f6e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.372081] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb24efcc-506f-448c-8748-826241b28f62 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.379705] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc2fcc5-fd04-4a4a-b375-feddd63c1fa6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.409382] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179837MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1302.409525] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1302.409727] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1302.602363] env[65385]: INFO nova.compute.manager [-] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Took 1.32 seconds to deallocate network for instance. [ 1303.107935] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1303.418116] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Applying migration context for instance 2c15094a-1c22-4e06-baf1-1be4a6fb432a as it has an incoming, in-progress migration 7448bd75-40bf-489b-b165-5ae8206b9677. Migration status is finished {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1303.419315] env[65385]: INFO nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating resource usage from migration 7448bd75-40bf-489b-b165-5ae8206b9677 [ 1303.439919] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance b4bca2c8-fc8f-471a-899b-aac648fee9fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1303.440093] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance c9f09e05-52c4-44fb-bfa8-85bde23ef8ce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1303.440217] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 120aa07c-d0c1-4031-8b61-986a261f2291 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1303.440333] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 6926f1bc-3c40-486f-b3e9-7e3731d57e4e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1303.440445] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1303.440559] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Migration 7448bd75-40bf-489b-b165-5ae8206b9677 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1303.440675] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 2c15094a-1c22-4e06-baf1-1be4a6fb432a actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1303.440882] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1303.441054] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=100GB used_disk=5GB total_vcpus=48 used_vcpus=7 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '6', 'num_vm_active': '5', 'num_task_None': '4', 'num_os_type_None': '6', 'num_proj_8d45ff3446a84ef2a1945f7a66b747e8': '3', 'io_workload': '0', 'num_vm_rescued': '1', 'num_proj_96214fc566ed4b11a92a6369a3419aea': '2', 'num_proj_df347ee234f142899014703c70ee42ed': '1', 'num_task_resize_finish': '1', 'num_task_deleting': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1303.536371] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc11e49f-b913-4f5c-bf53-bbd265f03c9d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.544475] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2d37f0-67db-44bd-9584-5a6e3af5186e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.575211] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4351d04-725d-43bd-be21-71c703c0e134 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.583954] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb7f6bd-6299-46db-85dc-a0b4a32be179 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.598692] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1304.102351] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1304.608820] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1304.609681] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.199s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1304.609681] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.502s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1304.609681] env[65385]: DEBUG nova.objects.instance [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lazy-loading 'resources' on Instance uuid 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1305.048891] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1305.049422] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1305.049734] env[65385]: DEBUG nova.compute.manager [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Going to confirm migration 10 {{(pid=65385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1305.230682] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79198191-a379-4d6b-8c03-788d09c8289e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.239715] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031debc2-9ce0-4e00-89b3-1ec26482b3d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.272906] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91fad5d-71ae-4f72-a756-3e7a5cd4eb3e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.281711] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77dff3e6-a95d-4c4d-b0d9-f257a1494324 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.297470] env[65385]: DEBUG nova.compute.provider_tree [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1305.557223] env[65385]: WARNING neutronclient.v2_0.client [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1305.594593] env[65385]: WARNING neutronclient.v2_0.client [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1305.595007] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1305.595177] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquired lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1305.595351] env[65385]: DEBUG nova.network.neutron [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1305.595536] env[65385]: DEBUG nova.objects.instance [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'info_cache' on Instance uuid 2c15094a-1c22-4e06-baf1-1be4a6fb432a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1305.610842] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1305.611199] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1305.611802] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1305.612030] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1305.801075] env[65385]: DEBUG nova.scheduler.client.report [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1306.306768] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1306.332596] env[65385]: INFO nova.scheduler.client.report [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Deleted allocations for instance 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb [ 1306.601823] env[65385]: WARNING neutronclient.v2_0.client [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1306.602563] env[65385]: WARNING openstack [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.602958] env[65385]: WARNING openstack [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1306.739170] env[65385]: WARNING openstack [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.739668] env[65385]: WARNING openstack [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1306.805320] env[65385]: WARNING neutronclient.v2_0.client [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1306.806015] env[65385]: WARNING openstack [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.806364] env[65385]: WARNING openstack [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1306.840816] env[65385]: DEBUG oslo_concurrency.lockutils [None req-61134757-c6db-4a13-8d3b-b4f6e0d5ccee tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "5d2820b9-db76-4572-b0ab-3c4a3ced0fdb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.180s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1306.887679] env[65385]: DEBUG nova.network.neutron [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [{"id": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "address": "fa:16:3e:57:7e:6a", "network": {"id": "d50f0f40-3598-4b4d-9bbc-270434bb79ac", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1699903063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d45ff3446a84ef2a1945f7a66b747e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "94e1d797-8eb2-4400-9f7d-f2eb60eb4cf2", "external-id": "nsx-vlan-transportzone-828", "segmentation_id": 828, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ce9de22-0e", "ovs_interfaceid": "5ce9de22-0e36-4aa4-86cc-4794bae3842c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1307.328549] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "120aa07c-d0c1-4031-8b61-986a261f2291" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1307.328814] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "120aa07c-d0c1-4031-8b61-986a261f2291" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1307.329032] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "120aa07c-d0c1-4031-8b61-986a261f2291-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1307.329224] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "120aa07c-d0c1-4031-8b61-986a261f2291-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1307.329392] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "120aa07c-d0c1-4031-8b61-986a261f2291-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1307.331485] env[65385]: INFO nova.compute.manager [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Terminating instance [ 1307.390878] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Releasing lock "refresh_cache-2c15094a-1c22-4e06-baf1-1be4a6fb432a" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1307.391157] env[65385]: DEBUG nova.objects.instance [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'migration_context' on Instance uuid 2c15094a-1c22-4e06-baf1-1be4a6fb432a {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1307.835332] env[65385]: DEBUG nova.compute.manager [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1307.835713] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1307.836473] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797c2b7b-370f-4c7d-ab69-08364df80b12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.845229] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.845500] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1307.845873] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8855887d-654c-49d5-a618-3ed449d3a12f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.853192] env[65385]: DEBUG oslo_vmware.api [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1307.853192] env[65385]: value = "task-4454751" [ 1307.853192] env[65385]: _type = "Task" [ 1307.853192] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.862763] env[65385]: DEBUG oslo_vmware.api [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454751, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.893862] env[65385]: DEBUG nova.objects.base [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Object Instance<2c15094a-1c22-4e06-baf1-1be4a6fb432a> lazy-loaded attributes: info_cache,migration_context {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1307.894953] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e69c6bb-e045-4208-9274-c18b03fa4b00 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.917713] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-132f4946-aba8-4268-8c61-7d2305528f02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.924554] env[65385]: DEBUG oslo_vmware.api [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1307.924554] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52762d7d-05af-2f22-dab1-5d8c11052384" [ 1307.924554] env[65385]: _type = "Task" [ 1307.924554] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.934706] env[65385]: DEBUG oslo_vmware.api [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52762d7d-05af-2f22-dab1-5d8c11052384, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.363107] env[65385]: DEBUG oslo_vmware.api [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454751, 'name': PowerOffVM_Task, 'duration_secs': 0.207511} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.364024] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1308.364024] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1308.364024] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30c3254a-aa43-4d92-85ca-1f074f02e808 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.433770] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1308.434150] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1308.434363] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Deleting the datastore file [datastore2] 120aa07c-d0c1-4031-8b61-986a261f2291 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1308.434731] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bcae116a-d2f8-405b-9cd4-7df9daea41ff {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.439974] env[65385]: DEBUG oslo_vmware.api [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52762d7d-05af-2f22-dab1-5d8c11052384, 'name': SearchDatastore_Task, 'duration_secs': 0.008862} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.440587] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1308.440879] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1308.445652] env[65385]: DEBUG oslo_vmware.api [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for the task: (returnval){ [ 1308.445652] env[65385]: value = "task-4454753" [ 1308.445652] env[65385]: _type = "Task" [ 1308.445652] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.454359] env[65385]: DEBUG oslo_vmware.api [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454753, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.955052] env[65385]: DEBUG oslo_vmware.api [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Task: {'id': task-4454753, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164022} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.955462] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1308.955462] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1308.955646] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1308.955840] env[65385]: INFO nova.compute.manager [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1308.956095] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1308.956295] env[65385]: DEBUG nova.compute.manager [-] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1308.956392] env[65385]: DEBUG nova.network.neutron [-] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1308.956636] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1308.957165] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1308.957415] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1308.996818] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1309.038208] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27435dd7-e667-4c8f-b14d-6dacd1fc212b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.046716] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36970745-801f-4d4b-a671-2bce76640c74 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.078239] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76f843c-c7e2-4bd4-9b1a-644a67d19086 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.086961] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69450b2d-d7a3-4b22-8efb-d4cccf28b3f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.101048] env[65385]: DEBUG nova.compute.provider_tree [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1309.258991] env[65385]: DEBUG nova.compute.manager [req-6d122200-75ef-41e8-977d-5e7b74631f4c req-ef9ad70a-a6b3-44eb-9a38-d923cd9e6b6b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Received event network-vif-deleted-0fcb3c6e-b016-42ef-aec2-92569a585155 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1309.259247] env[65385]: INFO nova.compute.manager [req-6d122200-75ef-41e8-977d-5e7b74631f4c req-ef9ad70a-a6b3-44eb-9a38-d923cd9e6b6b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Neutron deleted interface 0fcb3c6e-b016-42ef-aec2-92569a585155; detaching it from the instance and deleting it from the info cache [ 1309.259535] env[65385]: DEBUG nova.network.neutron [req-6d122200-75ef-41e8-977d-5e7b74631f4c req-ef9ad70a-a6b3-44eb-9a38-d923cd9e6b6b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1309.605347] env[65385]: DEBUG nova.scheduler.client.report [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1309.734638] env[65385]: DEBUG nova.network.neutron [-] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1309.762268] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5080d15b-ae62-4dbf-be11-735e240b91c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.772561] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17dfdea-81ad-43a6-9e20-b1731a78936e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.801919] env[65385]: DEBUG nova.compute.manager [req-6d122200-75ef-41e8-977d-5e7b74631f4c req-ef9ad70a-a6b3-44eb-9a38-d923cd9e6b6b service nova] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Detach interface failed, port_id=0fcb3c6e-b016-42ef-aec2-92569a585155, reason: Instance 120aa07c-d0c1-4031-8b61-986a261f2291 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1310.237604] env[65385]: INFO nova.compute.manager [-] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Took 1.28 seconds to deallocate network for instance. [ 1310.616799] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.176s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1310.744573] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1310.744920] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1310.745166] env[65385]: DEBUG nova.objects.instance [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lazy-loading 'resources' on Instance uuid 120aa07c-d0c1-4031-8b61-986a261f2291 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1311.178374] env[65385]: INFO nova.scheduler.client.report [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted allocation for migration 7448bd75-40bf-489b-b165-5ae8206b9677 [ 1311.328039] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2ef32a-2b6e-459b-8f60-26175b05be9c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.336471] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6e5721-03bf-4f14-a7fa-c67c3324e542 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.369656] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658ec93b-3dcb-4aad-a059-fc7aa031f87f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.378026] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da15ce2-086c-4ee8-a484-33ab59f9a14f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.391855] env[65385]: DEBUG nova.compute.provider_tree [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1311.540786] env[65385]: INFO nova.compute.manager [None req-d7b1ad9a-a040-406b-bbe6-02cd1da56855 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Get console output [ 1311.541527] env[65385]: WARNING nova.virt.vmwareapi.driver [None req-d7b1ad9a-a040-406b-bbe6-02cd1da56855 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] The console log is missing. Check your VSPC configuration [ 1311.685225] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fa210c3d-2570-4154-af1d-f950fae169f7 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.636s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1311.894883] env[65385]: DEBUG nova.scheduler.client.report [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1312.400422] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.655s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1312.421673] env[65385]: INFO nova.scheduler.client.report [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Deleted allocations for instance 120aa07c-d0c1-4031-8b61-986a261f2291 [ 1312.846486] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1312.846733] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1312.928843] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ce302880-1fe4-4342-ac84-3475120b81ed tempest-ServerRescueTestJSON-1506379550 tempest-ServerRescueTestJSON-1506379550-project-member] Lock "120aa07c-d0c1-4031-8b61-986a261f2291" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.600s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1313.349914] env[65385]: DEBUG nova.compute.utils [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1313.853221] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1314.922064] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1314.922064] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1314.922505] env[65385]: INFO nova.compute.manager [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Attaching volume 0e1ba855-b903-4512-8811-25211c190139 to /dev/sdb [ 1314.958524] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd8b2b7-23ed-4de5-a0b4-46a49363d348 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.967273] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb1520a-0a1b-44cf-acd1-d37b4a3e2584 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.984055] env[65385]: DEBUG nova.virt.block_device [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating existing volume attachment record: c37208a6-f173-4ac5-b8c1-c8a1982ba74b {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1318.675692] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "02e65e47-7971-4e08-acc9-eead1828d925" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1318.676120] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "02e65e47-7971-4e08-acc9-eead1828d925" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.020041] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.020220] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.179177] env[65385]: DEBUG nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1319.522804] env[65385]: DEBUG nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1319.530301] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1319.530555] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871218', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'name': 'volume-0e1ba855-b903-4512-8811-25211c190139', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6926f1bc-3c40-486f-b3e9-7e3731d57e4e', 'attached_at': '', 'detached_at': '', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'serial': '0e1ba855-b903-4512-8811-25211c190139'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1319.531417] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e9f054-8dca-48b4-a1d1-2943b555cfd0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.550639] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a108f2-aae9-4077-be06-4533f7f4bacd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.578133] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] volume-0e1ba855-b903-4512-8811-25211c190139/volume-0e1ba855-b903-4512-8811-25211c190139.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1319.578678] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5244561e-1f02-4fec-a456-1e70986c18a3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.598153] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1319.598153] env[65385]: value = "task-4454758" [ 1319.598153] env[65385]: _type = "Task" [ 1319.598153] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.607150] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454758, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.702254] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.702554] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.704367] env[65385]: INFO nova.compute.claims [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1320.042394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1320.110097] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454758, 'name': ReconfigVM_Task, 'duration_secs': 0.348669} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.110097] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Reconfigured VM instance instance-00000079 to attach disk [datastore2] volume-0e1ba855-b903-4512-8811-25211c190139/volume-0e1ba855-b903-4512-8811-25211c190139.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1320.113910] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f1d2021-e1b7-4ec6-a1d3-e352c624f27d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.129953] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1320.129953] env[65385]: value = "task-4454759" [ 1320.129953] env[65385]: _type = "Task" [ 1320.129953] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.138562] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454759, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.640481] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454759, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.798041] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f51e71-df42-44be-abad-bba74fbf32cf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.806799] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aafa33-ca3d-403a-b3b9-b92a4b312712 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.839419] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1465903e-7ac1-4a8a-a8f8-405b5e94384e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.847466] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d9e091-6058-4227-af7d-cfef404adda2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.861112] env[65385]: DEBUG nova.compute.provider_tree [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1321.141016] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454759, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.364340] env[65385]: DEBUG nova.scheduler.client.report [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1321.640984] env[65385]: DEBUG oslo_vmware.api [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454759, 'name': ReconfigVM_Task, 'duration_secs': 1.140761} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.641293] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871218', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'name': 'volume-0e1ba855-b903-4512-8811-25211c190139', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6926f1bc-3c40-486f-b3e9-7e3731d57e4e', 'attached_at': '', 'detached_at': '', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'serial': '0e1ba855-b903-4512-8811-25211c190139'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1321.869261] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.167s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1321.869801] env[65385]: DEBUG nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1321.872432] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.830s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1321.874265] env[65385]: INFO nova.compute.claims [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1322.382135] env[65385]: DEBUG nova.compute.utils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1322.384060] env[65385]: DEBUG nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1322.384222] env[65385]: DEBUG nova.network.neutron [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1322.384533] env[65385]: WARNING neutronclient.v2_0.client [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1322.384860] env[65385]: WARNING neutronclient.v2_0.client [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1322.385562] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1322.385962] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1322.452667] env[65385]: DEBUG nova.policy [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07886144152846b487b5d106416ee1bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fc382ad43724fd3ae1b9b62c1dd8131', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1322.677651] env[65385]: DEBUG nova.objects.instance [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1322.715607] env[65385]: DEBUG nova.network.neutron [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Successfully created port: 078edaec-5c64-4393-b4b9-0e9537ee060c {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1322.884434] env[65385]: DEBUG nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1322.982451] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec9dc5f-4f21-4cea-88c2-04d7f786bc8e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.991426] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54ff5bb-ca48-43e5-acc2-ef3678e9fffc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.022860] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44759dc4-80ee-499f-ad54-10b3cc1d15de {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.031146] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bc0643-e443-4294-9596-8d64ce9ea570 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.045110] env[65385]: DEBUG nova.compute.provider_tree [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1323.182205] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fb579098-7d6a-4226-a6fc-8fe4ff4fc2fd tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.260s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1323.351170] env[65385]: DEBUG oslo_concurrency.lockutils [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1323.351384] env[65385]: DEBUG oslo_concurrency.lockutils [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1323.351567] env[65385]: DEBUG nova.compute.manager [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1323.352609] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccb96b4-6e3b-43ac-8241-d6a7bc10d193 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.362278] env[65385]: DEBUG nova.compute.manager [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1323.362890] env[65385]: DEBUG nova.objects.instance [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1323.547996] env[65385]: DEBUG nova.scheduler.client.report [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1323.895768] env[65385]: DEBUG nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1323.921352] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1323.921623] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1323.921775] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1323.921949] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1323.922102] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1323.922243] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1323.922442] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1323.922591] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1323.922751] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1323.922901] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1323.923090] env[65385]: DEBUG nova.virt.hardware [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1323.923944] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9cf55c4-895d-49ff-86e2-339f732a0a88 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.932162] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa7e333-a972-4958-98b4-71eaa199c680 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.052581] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.180s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1324.053145] env[65385]: DEBUG nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1324.137597] env[65385]: DEBUG nova.compute.manager [req-f1b63a6e-8117-4a58-a3d3-263b12b4479f req-5a070bab-0697-4b8e-b3ab-fcaa98e4381f service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Received event network-vif-plugged-078edaec-5c64-4393-b4b9-0e9537ee060c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1324.137805] env[65385]: DEBUG oslo_concurrency.lockutils [req-f1b63a6e-8117-4a58-a3d3-263b12b4479f req-5a070bab-0697-4b8e-b3ab-fcaa98e4381f service nova] Acquiring lock "02e65e47-7971-4e08-acc9-eead1828d925-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1324.138016] env[65385]: DEBUG oslo_concurrency.lockutils [req-f1b63a6e-8117-4a58-a3d3-263b12b4479f req-5a070bab-0697-4b8e-b3ab-fcaa98e4381f service nova] Lock "02e65e47-7971-4e08-acc9-eead1828d925-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1324.138205] env[65385]: DEBUG oslo_concurrency.lockutils [req-f1b63a6e-8117-4a58-a3d3-263b12b4479f req-5a070bab-0697-4b8e-b3ab-fcaa98e4381f service nova] Lock "02e65e47-7971-4e08-acc9-eead1828d925-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1324.138340] env[65385]: DEBUG nova.compute.manager [req-f1b63a6e-8117-4a58-a3d3-263b12b4479f req-5a070bab-0697-4b8e-b3ab-fcaa98e4381f service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] No waiting events found dispatching network-vif-plugged-078edaec-5c64-4393-b4b9-0e9537ee060c {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1324.138501] env[65385]: WARNING nova.compute.manager [req-f1b63a6e-8117-4a58-a3d3-263b12b4479f req-5a070bab-0697-4b8e-b3ab-fcaa98e4381f service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Received unexpected event network-vif-plugged-078edaec-5c64-4393-b4b9-0e9537ee060c for instance with vm_state building and task_state spawning. [ 1324.225048] env[65385]: DEBUG nova.network.neutron [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Successfully updated port: 078edaec-5c64-4393-b4b9-0e9537ee060c {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1324.369478] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1324.369806] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-726b89cc-5536-403d-bac6-615648ce58eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.377620] env[65385]: DEBUG oslo_vmware.api [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1324.377620] env[65385]: value = "task-4454760" [ 1324.377620] env[65385]: _type = "Task" [ 1324.377620] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.386362] env[65385]: DEBUG oslo_vmware.api [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.558544] env[65385]: DEBUG nova.compute.utils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1324.560953] env[65385]: DEBUG nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1324.561388] env[65385]: DEBUG nova.network.neutron [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1324.561611] env[65385]: WARNING neutronclient.v2_0.client [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1324.562013] env[65385]: WARNING neutronclient.v2_0.client [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1324.562618] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1324.563015] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1324.608046] env[65385]: DEBUG nova.policy [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07886144152846b487b5d106416ee1bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fc382ad43724fd3ae1b9b62c1dd8131', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1324.727610] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1324.727857] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1324.727971] env[65385]: DEBUG nova.network.neutron [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1324.879453] env[65385]: DEBUG nova.network.neutron [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Successfully created port: 764d3f51-03a5-48ff-bab9-f005c7613395 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1324.891641] env[65385]: DEBUG oslo_vmware.api [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454760, 'name': PowerOffVM_Task, 'duration_secs': 0.191365} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.891920] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1324.892138] env[65385]: DEBUG nova.compute.manager [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1324.892929] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05407a9e-b68a-486f-9d57-3fccfea2ccc3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.061581] env[65385]: DEBUG nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1325.232027] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1325.232027] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1325.265606] env[65385]: DEBUG nova.network.neutron [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1325.285672] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1325.286221] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1325.350616] env[65385]: WARNING neutronclient.v2_0.client [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.351386] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1325.351737] env[65385]: WARNING openstack [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1325.405383] env[65385]: DEBUG oslo_concurrency.lockutils [None req-efc06079-b6ba-4826-a8f7-9dbca4bff15a tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1325.463574] env[65385]: DEBUG nova.network.neutron [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Updating instance_info_cache with network_info: [{"id": "078edaec-5c64-4393-b4b9-0e9537ee060c", "address": "fa:16:3e:62:87:51", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap078edaec-5c", "ovs_interfaceid": "078edaec-5c64-4393-b4b9-0e9537ee060c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1325.797140] env[65385]: DEBUG nova.objects.instance [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1325.965502] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1325.966075] env[65385]: DEBUG nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Instance network_info: |[{"id": "078edaec-5c64-4393-b4b9-0e9537ee060c", "address": "fa:16:3e:62:87:51", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap078edaec-5c", "ovs_interfaceid": "078edaec-5c64-4393-b4b9-0e9537ee060c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1325.966358] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:87:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1559ce49-7345-443f-bf02-4bfeb88356ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '078edaec-5c64-4393-b4b9-0e9537ee060c', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1325.973929] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating folder: Project (2fc382ad43724fd3ae1b9b62c1dd8131). Parent ref: group-v870881. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1325.974243] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90516218-6d63-4660-a0ad-7a3dde586782 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.987054] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created folder: Project (2fc382ad43724fd3ae1b9b62c1dd8131) in parent group-v870881. [ 1325.987267] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating folder: Instances. Parent ref: group-v871219. {{(pid=65385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1325.987510] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82604619-5c2c-4fa4-9a6f-6c7e5f11707e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.997914] env[65385]: INFO nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created folder: Instances in parent group-v871219. [ 1325.998203] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1325.998403] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1325.998613] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-adfe9c45-9208-46f0-8ea8-a3e61c1d3a30 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.019569] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1326.019569] env[65385]: value = "task-4454763" [ 1326.019569] env[65385]: _type = "Task" [ 1326.019569] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.027827] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454763, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.071687] env[65385]: DEBUG nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1326.101805] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1326.102063] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1326.102222] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1326.102398] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1326.102535] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1326.102672] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1326.102899] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1326.103061] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1326.103226] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1326.103380] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1326.103541] env[65385]: DEBUG nova.virt.hardware [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1326.104438] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c050190-4563-42d6-a6c0-401b1011523a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.113328] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08479ebb-b256-4557-8d2c-c0deebdd444c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.166705] env[65385]: DEBUG nova.compute.manager [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Received event network-changed-078edaec-5c64-4393-b4b9-0e9537ee060c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1326.166916] env[65385]: DEBUG nova.compute.manager [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Refreshing instance network info cache due to event network-changed-078edaec-5c64-4393-b4b9-0e9537ee060c. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1326.167164] env[65385]: DEBUG oslo_concurrency.lockutils [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Acquiring lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.167345] env[65385]: DEBUG oslo_concurrency.lockutils [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Acquired lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1326.167521] env[65385]: DEBUG nova.network.neutron [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Refreshing network info cache for port 078edaec-5c64-4393-b4b9-0e9537ee060c {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1326.293196] env[65385]: DEBUG nova.compute.manager [req-6029c145-a4f4-4d4b-819f-ac6d016c14da req-7adb03bd-e7e3-4be5-a10f-12c872044fd1 service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Received event network-vif-plugged-764d3f51-03a5-48ff-bab9-f005c7613395 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1326.293439] env[65385]: DEBUG oslo_concurrency.lockutils [req-6029c145-a4f4-4d4b-819f-ac6d016c14da req-7adb03bd-e7e3-4be5-a10f-12c872044fd1 service nova] Acquiring lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1326.293694] env[65385]: DEBUG oslo_concurrency.lockutils [req-6029c145-a4f4-4d4b-819f-ac6d016c14da req-7adb03bd-e7e3-4be5-a10f-12c872044fd1 service nova] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1326.293883] env[65385]: DEBUG oslo_concurrency.lockutils [req-6029c145-a4f4-4d4b-819f-ac6d016c14da req-7adb03bd-e7e3-4be5-a10f-12c872044fd1 service nova] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1326.294119] env[65385]: DEBUG nova.compute.manager [req-6029c145-a4f4-4d4b-819f-ac6d016c14da req-7adb03bd-e7e3-4be5-a10f-12c872044fd1 service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] No waiting events found dispatching network-vif-plugged-764d3f51-03a5-48ff-bab9-f005c7613395 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1326.294296] env[65385]: WARNING nova.compute.manager [req-6029c145-a4f4-4d4b-819f-ac6d016c14da req-7adb03bd-e7e3-4be5-a10f-12c872044fd1 service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Received unexpected event network-vif-plugged-764d3f51-03a5-48ff-bab9-f005c7613395 for instance with vm_state building and task_state spawning. [ 1326.303157] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.303477] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1326.303793] env[65385]: DEBUG nova.network.neutron [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1326.304081] env[65385]: DEBUG nova.objects.instance [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'info_cache' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1326.398265] env[65385]: DEBUG nova.network.neutron [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Successfully updated port: 764d3f51-03a5-48ff-bab9-f005c7613395 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1326.530087] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454763, 'name': CreateVM_Task, 'duration_secs': 0.346814} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.530279] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1326.530853] env[65385]: WARNING neutronclient.v2_0.client [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1326.531346] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.531423] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1326.531828] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1326.532046] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4435e957-1867-4f90-be96-693ede08bbea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.537263] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1326.537263] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5229cb87-b0ac-c3ea-24ad-088347b49dc9" [ 1326.537263] env[65385]: _type = "Task" [ 1326.537263] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.546054] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5229cb87-b0ac-c3ea-24ad-088347b49dc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.669791] env[65385]: WARNING neutronclient.v2_0.client [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1326.670989] env[65385]: WARNING openstack [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1326.670989] env[65385]: WARNING openstack [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1326.765253] env[65385]: WARNING openstack [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1326.765697] env[65385]: WARNING openstack [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1326.807554] env[65385]: DEBUG nova.objects.base [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Object Instance<6926f1bc-3c40-486f-b3e9-7e3731d57e4e> lazy-loaded attributes: flavor,info_cache {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1326.822492] env[65385]: WARNING neutronclient.v2_0.client [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1326.823112] env[65385]: WARNING openstack [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1326.823462] env[65385]: WARNING openstack [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1326.900681] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.900871] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1326.901083] env[65385]: DEBUG nova.network.neutron [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1326.905767] env[65385]: DEBUG nova.network.neutron [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Updated VIF entry in instance network info cache for port 078edaec-5c64-4393-b4b9-0e9537ee060c. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1326.906198] env[65385]: DEBUG nova.network.neutron [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Updating instance_info_cache with network_info: [{"id": "078edaec-5c64-4393-b4b9-0e9537ee060c", "address": "fa:16:3e:62:87:51", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap078edaec-5c", "ovs_interfaceid": "078edaec-5c64-4393-b4b9-0e9537ee060c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1327.051235] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5229cb87-b0ac-c3ea-24ad-088347b49dc9, 'name': SearchDatastore_Task, 'duration_secs': 0.013279} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.051551] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1327.051811] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1327.052075] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1327.052219] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1327.052499] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1327.052810] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a94ea20-d5f8-4f81-af22-8d6bc9c95708 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.062532] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1327.062696] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1327.063454] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22964b2d-27f5-4f07-b25f-feaf6f5521fe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.069322] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1327.069322] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a7a564-375a-9e1a-85e3-0f5e975247e1" [ 1327.069322] env[65385]: _type = "Task" [ 1327.069322] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.077489] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a7a564-375a-9e1a-85e3-0f5e975247e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.310116] env[65385]: WARNING neutronclient.v2_0.client [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1327.310847] env[65385]: WARNING openstack [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1327.311216] env[65385]: WARNING openstack [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1327.403939] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1327.404374] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1327.411610] env[65385]: DEBUG oslo_concurrency.lockutils [req-10ada121-78f0-40df-8a1e-411b7198476d req-80d27fd7-7bbe-4d31-9624-3aaa2e0126d3 service nova] Releasing lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1327.419033] env[65385]: WARNING openstack [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1327.419033] env[65385]: WARNING openstack [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1327.438012] env[65385]: DEBUG nova.network.neutron [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1327.454429] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1327.454810] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1327.472660] env[65385]: WARNING neutronclient.v2_0.client [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1327.473287] env[65385]: WARNING openstack [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1327.473625] env[65385]: WARNING openstack [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1327.510041] env[65385]: WARNING neutronclient.v2_0.client [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1327.510703] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1327.511013] env[65385]: WARNING openstack [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1327.556393] env[65385]: DEBUG nova.network.neutron [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating instance_info_cache with network_info: [{"id": "59e7b718-2455-470d-85a3-3a0d0849ea44", "address": "fa:16:3e:7e:65:d9", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e7b718-24", "ovs_interfaceid": "59e7b718-2455-470d-85a3-3a0d0849ea44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1327.581147] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52a7a564-375a-9e1a-85e3-0f5e975247e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009048} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.581950] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b72d357-0ed9-48ef-891d-203e7b026951 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.588368] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1327.588368] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52966e3b-d5b9-94d7-4a42-883ac3fb6cbe" [ 1327.588368] env[65385]: _type = "Task" [ 1327.588368] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.594153] env[65385]: DEBUG nova.network.neutron [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Updating instance_info_cache with network_info: [{"id": "764d3f51-03a5-48ff-bab9-f005c7613395", "address": "fa:16:3e:5e:3d:46", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764d3f51-03", "ovs_interfaceid": "764d3f51-03a5-48ff-bab9-f005c7613395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1327.601276] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52966e3b-d5b9-94d7-4a42-883ac3fb6cbe, 'name': SearchDatastore_Task, 'duration_secs': 0.010259} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.601500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1327.601733] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/02e65e47-7971-4e08-acc9-eead1828d925.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1327.601976] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-077735dd-4801-4796-ae78-706d4a8624ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.609082] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1327.609082] env[65385]: value = "task-4454764" [ 1327.609082] env[65385]: _type = "Task" [ 1327.609082] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.617929] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454764, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.059725] env[65385]: DEBUG oslo_concurrency.lockutils [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1328.098295] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1328.098753] env[65385]: DEBUG nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Instance network_info: |[{"id": "764d3f51-03a5-48ff-bab9-f005c7613395", "address": "fa:16:3e:5e:3d:46", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764d3f51-03", "ovs_interfaceid": "764d3f51-03a5-48ff-bab9-f005c7613395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1328.099160] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:3d:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1559ce49-7345-443f-bf02-4bfeb88356ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '764d3f51-03a5-48ff-bab9-f005c7613395', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1328.106619] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1328.107068] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1328.107331] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efda2594-3e6a-4859-96ea-d448b1c65fc6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.132238] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454764, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441515} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.133592] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/02e65e47-7971-4e08-acc9-eead1828d925.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1328.133850] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1328.134122] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1328.134122] env[65385]: value = "task-4454765" [ 1328.134122] env[65385]: _type = "Task" [ 1328.134122] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.134323] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18227d24-c8fe-4cff-b601-148673bf093d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.145054] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454765, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.145996] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1328.145996] env[65385]: value = "task-4454766" [ 1328.145996] env[65385]: _type = "Task" [ 1328.145996] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.155525] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.322498] env[65385]: DEBUG nova.compute.manager [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Received event network-changed-764d3f51-03a5-48ff-bab9-f005c7613395 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1328.322498] env[65385]: DEBUG nova.compute.manager [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Refreshing instance network info cache due to event network-changed-764d3f51-03a5-48ff-bab9-f005c7613395. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1328.322740] env[65385]: DEBUG oslo_concurrency.lockutils [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Acquiring lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.322833] env[65385]: DEBUG oslo_concurrency.lockutils [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Acquired lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1328.322987] env[65385]: DEBUG nova.network.neutron [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Refreshing network info cache for port 764d3f51-03a5-48ff-bab9-f005c7613395 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1328.646682] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454765, 'name': CreateVM_Task, 'duration_secs': 0.422093} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.649676] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1328.650205] env[65385]: WARNING neutronclient.v2_0.client [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.650565] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.650710] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1328.651048] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1328.651657] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3717e916-b9b0-454d-9eb8-1bfcc03435a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.657268] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07035} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.658555] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1328.658895] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1328.658895] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52347cda-a4dd-0c5d-2fa6-00967c74be6a" [ 1328.658895] env[65385]: _type = "Task" [ 1328.658895] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.659594] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af8ab5b-1706-4484-a923-b6f63f886a7f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.670847] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52347cda-a4dd-0c5d-2fa6-00967c74be6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.688271] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/02e65e47-7971-4e08-acc9-eead1828d925.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1328.688630] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e87d6194-7950-4b45-8b53-45f95390eda2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.710229] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1328.710229] env[65385]: value = "task-4454767" [ 1328.710229] env[65385]: _type = "Task" [ 1328.710229] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.719576] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454767, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.825670] env[65385]: WARNING neutronclient.v2_0.client [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.826559] env[65385]: WARNING openstack [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.826923] env[65385]: WARNING openstack [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.921721] env[65385]: WARNING openstack [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.922170] env[65385]: WARNING openstack [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.985683] env[65385]: WARNING neutronclient.v2_0.client [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.986436] env[65385]: WARNING openstack [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.986821] env[65385]: WARNING openstack [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1329.066609] env[65385]: DEBUG nova.network.neutron [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Updated VIF entry in instance network info cache for port 764d3f51-03a5-48ff-bab9-f005c7613395. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1329.066998] env[65385]: DEBUG nova.network.neutron [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Updating instance_info_cache with network_info: [{"id": "764d3f51-03a5-48ff-bab9-f005c7613395", "address": "fa:16:3e:5e:3d:46", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764d3f51-03", "ovs_interfaceid": "764d3f51-03a5-48ff-bab9-f005c7613395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1329.068631] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1329.069104] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c02f520c-668d-4d90-9daf-2f54e82327f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.079070] env[65385]: DEBUG oslo_vmware.api [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1329.079070] env[65385]: value = "task-4454768" [ 1329.079070] env[65385]: _type = "Task" [ 1329.079070] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.088241] env[65385]: DEBUG oslo_vmware.api [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454768, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.174509] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52347cda-a4dd-0c5d-2fa6-00967c74be6a, 'name': SearchDatastore_Task, 'duration_secs': 0.01267} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.174872] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1329.175052] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1329.175303] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1329.175441] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1329.175605] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1329.175861] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8f4bb45-acc8-4dd9-bb1c-50d7b69ea6ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.184954] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1329.185148] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1329.185878] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f659cff3-1495-4e37-9379-c54dfdf1c88c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.191334] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1329.191334] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b7be6e-5239-2ea2-4b04-73ee7f31d9c9" [ 1329.191334] env[65385]: _type = "Task" [ 1329.191334] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.199557] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b7be6e-5239-2ea2-4b04-73ee7f31d9c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.220164] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454767, 'name': ReconfigVM_Task, 'duration_secs': 0.293053} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.220431] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Reconfigured VM instance instance-0000007c to attach disk [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/02e65e47-7971-4e08-acc9-eead1828d925.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1329.221059] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d2f75a4-857e-4b68-9cfe-123b1824c2a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.228062] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1329.228062] env[65385]: value = "task-4454769" [ 1329.228062] env[65385]: _type = "Task" [ 1329.228062] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.239846] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454769, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.570135] env[65385]: DEBUG oslo_concurrency.lockutils [req-40c58c8d-9b5a-4e4d-a14c-958f8f54ae8b req-e0dc068a-58a5-4b12-a19d-607aebfdd27e service nova] Releasing lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1329.589520] env[65385]: DEBUG oslo_vmware.api [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454768, 'name': PowerOnVM_Task, 'duration_secs': 0.41085} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.589789] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1329.589989] env[65385]: DEBUG nova.compute.manager [None req-bb339c6d-13e9-43ae-84da-a3c1ba004abf tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1329.590801] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31cbca39-b5d3-42dd-8468-636096551853 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.702152] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b7be6e-5239-2ea2-4b04-73ee7f31d9c9, 'name': SearchDatastore_Task, 'duration_secs': 0.008939} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.703107] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-527074e2-d937-48b2-a45c-134159cbd1e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.708280] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1329.708280] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f51473-46c8-a320-13da-f0333085f718" [ 1329.708280] env[65385]: _type = "Task" [ 1329.708280] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.715916] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f51473-46c8-a320-13da-f0333085f718, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.735461] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454769, 'name': Rename_Task, 'duration_secs': 0.204303} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.735695] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1329.735922] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2cbb917-21f0-4e46-9844-e924b541ab8e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.743110] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1329.743110] env[65385]: value = "task-4454770" [ 1329.743110] env[65385]: _type = "Task" [ 1329.743110] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.750857] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.219123] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52f51473-46c8-a320-13da-f0333085f718, 'name': SearchDatastore_Task, 'duration_secs': 0.0096} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.219500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1330.219622] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/ad85aa5c-68a6-41a3-88d7-c27bbef81c6f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1330.219879] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58a36fef-d257-45dd-972a-56206f8f9d20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.227153] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1330.227153] env[65385]: value = "task-4454771" [ 1330.227153] env[65385]: _type = "Task" [ 1330.227153] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.235617] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.254213] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454770, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.737422] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455335} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.737612] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/ad85aa5c-68a6-41a3-88d7-c27bbef81c6f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1330.737815] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1330.738080] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a56faa1-946f-4e8b-9a66-ba600926fee1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.745195] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1330.745195] env[65385]: value = "task-4454772" [ 1330.745195] env[65385]: _type = "Task" [ 1330.745195] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.757291] env[65385]: DEBUG oslo_vmware.api [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454770, 'name': PowerOnVM_Task, 'duration_secs': 0.588158} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.760284] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1330.760498] env[65385]: INFO nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Took 6.86 seconds to spawn the instance on the hypervisor. [ 1330.760679] env[65385]: DEBUG nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1330.760965] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.762045] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6501ee5-0ed3-426f-b619-ec3898633e08 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.260771] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073794} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.261171] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1331.261861] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923d6762-bd76-490d-bb90-e1c5be5261d2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.287936] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/ad85aa5c-68a6-41a3-88d7-c27bbef81c6f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1331.289869] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8470580-e4b5-4b27-9e1d-80ae4f215dcd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.304688] env[65385]: INFO nova.compute.manager [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Took 11.62 seconds to build instance. [ 1331.312586] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1331.312586] env[65385]: value = "task-4454773" [ 1331.312586] env[65385]: _type = "Task" [ 1331.312586] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.322283] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454773, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.807061] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e8206a4e-ffb1-4874-ae64-08ff4e9bea7f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "02e65e47-7971-4e08-acc9-eead1828d925" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.131s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1331.823393] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454773, 'name': ReconfigVM_Task, 'duration_secs': 0.318061} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.823650] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Reconfigured VM instance instance-0000007d to attach disk [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/ad85aa5c-68a6-41a3-88d7-c27bbef81c6f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1331.824271] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6d55120-4993-4a72-aeb6-412525dac330 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.831059] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1331.831059] env[65385]: value = "task-4454774" [ 1331.831059] env[65385]: _type = "Task" [ 1331.831059] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.840812] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454774, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.341114] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454774, 'name': Rename_Task, 'duration_secs': 0.187988} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.341484] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1332.341622] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4871dc4-7a38-4707-9999-b5edae2fcfa1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.348593] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1332.348593] env[65385]: value = "task-4454775" [ 1332.348593] env[65385]: _type = "Task" [ 1332.348593] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.357105] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.859605] env[65385]: DEBUG oslo_vmware.api [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454775, 'name': PowerOnVM_Task, 'duration_secs': 0.470138} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.859842] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1332.860014] env[65385]: INFO nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Took 6.79 seconds to spawn the instance on the hypervisor. [ 1332.860196] env[65385]: DEBUG nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1332.860976] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7936c18-6790-410b-abae-bcee4caeb920 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.378321] env[65385]: INFO nova.compute.manager [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Took 13.35 seconds to build instance. [ 1333.880266] env[65385]: DEBUG oslo_concurrency.lockutils [None req-987734cf-267f-4d80-a229-1160e60ac005 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.860s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1334.351334] env[65385]: INFO nova.compute.manager [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Rescuing [ 1334.351637] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1334.351757] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1334.351897] env[65385]: DEBUG nova.network.neutron [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1334.858110] env[65385]: WARNING neutronclient.v2_0.client [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1334.858110] env[65385]: WARNING openstack [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1334.858543] env[65385]: WARNING openstack [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1334.977513] env[65385]: WARNING openstack [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1334.977513] env[65385]: WARNING openstack [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1335.036188] env[65385]: WARNING neutronclient.v2_0.client [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1335.036836] env[65385]: WARNING openstack [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1335.037312] env[65385]: WARNING openstack [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1335.116684] env[65385]: DEBUG nova.network.neutron [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Updating instance_info_cache with network_info: [{"id": "764d3f51-03a5-48ff-bab9-f005c7613395", "address": "fa:16:3e:5e:3d:46", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764d3f51-03", "ovs_interfaceid": "764d3f51-03a5-48ff-bab9-f005c7613395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1335.620325] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1337.161728] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1337.162312] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ce5dc9c-2053-4fe8-a1f3-f94f27dc55f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.170206] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1337.170206] env[65385]: value = "task-4454776" [ 1337.170206] env[65385]: _type = "Task" [ 1337.170206] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.181438] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.681098] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454776, 'name': PowerOffVM_Task, 'duration_secs': 0.187564} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.681324] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1337.682111] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9c20b4-87da-4997-9d26-b7b0251419dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.700515] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fbc955-4177-4adb-a4ce-560060a733d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.733152] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1337.733431] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55feed57-1c26-4411-a6c3-2a1cb34d4a1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.740642] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1337.740642] env[65385]: value = "task-4454777" [ 1337.740642] env[65385]: _type = "Task" [ 1337.740642] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.748677] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.033943] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1338.034275] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1338.034492] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1338.034672] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1338.034834] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1338.037022] env[65385]: INFO nova.compute.manager [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Terminating instance [ 1338.252059] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1338.252457] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1338.252505] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1338.252652] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1338.252825] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1338.253130] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c81b66cd-6c99-468f-817a-4e9722939347 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.267130] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1338.267349] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1338.268311] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d692b221-3baf-4ab8-bba0-d8e43ab0fda0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.276145] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1338.276145] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525078fb-f7d2-9974-76af-72ee85040d57" [ 1338.276145] env[65385]: _type = "Task" [ 1338.276145] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.284642] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525078fb-f7d2-9974-76af-72ee85040d57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.542633] env[65385]: DEBUG nova.compute.manager [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1338.542826] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1338.543163] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43df948b-4a8c-448d-959d-f59e4d5de2a6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.550605] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1338.550605] env[65385]: value = "task-4454778" [ 1338.550605] env[65385]: _type = "Task" [ 1338.550605] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.560211] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.786799] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]525078fb-f7d2-9974-76af-72ee85040d57, 'name': SearchDatastore_Task, 'duration_secs': 0.010322} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.787750] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cc43428-93e4-478c-9352-081ba5b652f2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.793395] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1338.793395] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eac4c1-9366-994b-df8a-88b0e1e1d033" [ 1338.793395] env[65385]: _type = "Task" [ 1338.793395] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.801449] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eac4c1-9366-994b-df8a-88b0e1e1d033, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.061515] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454778, 'name': PowerOffVM_Task, 'duration_secs': 0.165615} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.061794] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1339.062073] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1339.062273] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871211', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'name': 'volume-abe0b31f-295c-4222-ba80-83d0bae39141', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '2c15094a-1c22-4e06-baf1-1be4a6fb432a', 'attached_at': '2025-11-14T16:58:59.000000', 'detached_at': '', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'serial': 'abe0b31f-295c-4222-ba80-83d0bae39141'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1339.063067] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774ac34b-f174-42fd-82f1-80b51d3f8de6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.082838] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f016b3e-cd83-4564-9c5a-870496afbf09 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.090778] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c351392-3f71-4ded-95b3-e470c04ae13d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.109023] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effacf08-9d9c-4764-9407-eeaf50fc9687 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.125774] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] The volume has not been displaced from its original location: [datastore1] volume-abe0b31f-295c-4222-ba80-83d0bae39141/volume-abe0b31f-295c-4222-ba80-83d0bae39141.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1339.130768] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfiguring VM instance instance-0000007a to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1339.131071] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-263b0ec2-833f-4a36-b382-fc17eb9ca486 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.149558] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1339.149558] env[65385]: value = "task-4454779" [ 1339.149558] env[65385]: _type = "Task" [ 1339.149558] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.158254] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454779, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.303839] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52eac4c1-9366-994b-df8a-88b0e1e1d033, 'name': SearchDatastore_Task, 'duration_secs': 0.010276} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.304233] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1339.304361] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. {{(pid=65385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1339.304641] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-213f82f2-c33d-40c2-a81b-c6fcdedd516a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.313441] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1339.313441] env[65385]: value = "task-4454780" [ 1339.313441] env[65385]: _type = "Task" [ 1339.313441] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.322065] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454780, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.663037] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454779, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.824706] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454780, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.160702] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454779, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.330378] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454780, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645381} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.330748] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. [ 1340.331644] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8524cd71-2cea-4423-b922-26f3e2a8802b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.357054] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1340.357371] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-112d93f4-6472-4d78-acf1-d9674764d66d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.375885] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1340.375885] env[65385]: value = "task-4454781" [ 1340.375885] env[65385]: _type = "Task" [ 1340.375885] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.384391] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454781, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.661333] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454779, 'name': ReconfigVM_Task, 'duration_secs': 1.204313} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.661624] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Reconfigured VM instance instance-0000007a to detach disk 2000 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1340.666428] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48df9fde-6d92-4890-b7a7-30422c4106e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.681970] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1340.681970] env[65385]: value = "task-4454782" [ 1340.681970] env[65385]: _type = "Task" [ 1340.681970] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.692107] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454782, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.885999] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454781, 'name': ReconfigVM_Task, 'duration_secs': 0.313366} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.886361] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Reconfigured VM instance instance-0000007d to attach disk [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1340.887273] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc75f56-9e98-48e2-b760-327e9dbe141d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.913685] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be7cb99a-5477-4a34-a250-5178edc26761 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.930277] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1340.930277] env[65385]: value = "task-4454783" [ 1340.930277] env[65385]: _type = "Task" [ 1340.930277] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.939175] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454783, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.192704] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454782, 'name': ReconfigVM_Task, 'duration_secs': 0.113141} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.192994] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871211', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'name': 'volume-abe0b31f-295c-4222-ba80-83d0bae39141', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '2c15094a-1c22-4e06-baf1-1be4a6fb432a', 'attached_at': '2025-11-14T16:58:59.000000', 'detached_at': '', 'volume_id': 'abe0b31f-295c-4222-ba80-83d0bae39141', 'serial': 'abe0b31f-295c-4222-ba80-83d0bae39141'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1341.193323] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1341.194107] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825e8765-6170-4a71-a4ed-00348ea339da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.201290] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1341.201524] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6343547b-ffd1-44e2-ace3-2591068dfbc2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.271268] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1341.271537] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1341.271751] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleting the datastore file [datastore1] 2c15094a-1c22-4e06-baf1-1be4a6fb432a {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1341.272087] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27e0effe-77f5-4754-ac82-778f66bfe414 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.279621] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1341.279621] env[65385]: value = "task-4454785" [ 1341.279621] env[65385]: _type = "Task" [ 1341.279621] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.288704] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.440860] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454783, 'name': ReconfigVM_Task, 'duration_secs': 0.166543} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.441214] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1341.441356] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-984d48fb-2146-495d-a968-4c472655b971 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.448759] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1341.448759] env[65385]: value = "task-4454786" [ 1341.448759] env[65385]: _type = "Task" [ 1341.448759] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.457288] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454786, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.790761] env[65385]: DEBUG oslo_vmware.api [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.318447} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.791045] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1341.791237] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1341.791422] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1341.791590] env[65385]: INFO nova.compute.manager [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Took 3.25 seconds to destroy the instance on the hypervisor. [ 1341.791828] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1341.792036] env[65385]: DEBUG nova.compute.manager [-] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1341.792137] env[65385]: DEBUG nova.network.neutron [-] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1341.792408] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1341.792954] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1341.793248] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1341.836338] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1341.959827] env[65385]: DEBUG oslo_vmware.api [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454786, 'name': PowerOnVM_Task, 'duration_secs': 0.397701} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.960014] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1341.962854] env[65385]: DEBUG nova.compute.manager [None req-b00d10b7-05a8-4ee5-9307-d7e0a96a9641 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1341.963697] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12459be-3ace-40db-999c-d5fd2490d951 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.294641] env[65385]: DEBUG nova.compute.manager [req-ee901c82-e566-4263-ae14-e1fb352c3144 req-1cce92c9-e814-4bf7-b100-58593f828118 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Received event network-vif-deleted-5ce9de22-0e36-4aa4-86cc-4794bae3842c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1342.294871] env[65385]: INFO nova.compute.manager [req-ee901c82-e566-4263-ae14-e1fb352c3144 req-1cce92c9-e814-4bf7-b100-58593f828118 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Neutron deleted interface 5ce9de22-0e36-4aa4-86cc-4794bae3842c; detaching it from the instance and deleting it from the info cache [ 1342.295049] env[65385]: DEBUG nova.network.neutron [req-ee901c82-e566-4263-ae14-e1fb352c3144 req-1cce92c9-e814-4bf7-b100-58593f828118 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1342.768848] env[65385]: DEBUG nova.network.neutron [-] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1342.798449] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-957cb267-fe5f-4f43-8ac5-c94485484fde {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.808870] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8ca232-45a8-4bbb-b5c8-825e96cea433 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.841103] env[65385]: DEBUG nova.compute.manager [req-ee901c82-e566-4263-ae14-e1fb352c3144 req-1cce92c9-e814-4bf7-b100-58593f828118 service nova] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Detach interface failed, port_id=5ce9de22-0e36-4aa4-86cc-4794bae3842c, reason: Instance 2c15094a-1c22-4e06-baf1-1be4a6fb432a could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1343.272220] env[65385]: INFO nova.compute.manager [-] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Took 1.48 seconds to deallocate network for instance. [ 1343.815767] env[65385]: INFO nova.compute.manager [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Took 0.54 seconds to detach 1 volumes for instance. [ 1343.817929] env[65385]: DEBUG nova.compute.manager [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Deleting volume: abe0b31f-295c-4222-ba80-83d0bae39141 {{(pid=65385) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3320}} [ 1344.352021] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1344.352321] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1344.352450] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1344.378065] env[65385]: INFO nova.scheduler.client.report [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted allocations for instance 2c15094a-1c22-4e06-baf1-1be4a6fb432a [ 1344.863930] env[65385]: INFO nova.compute.manager [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Rescuing [ 1344.864337] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1344.864435] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1344.864637] env[65385]: DEBUG nova.network.neutron [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1344.885229] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b9d23ed6-fe2e-4c09-8325-806ba5c7d56b tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "2c15094a-1c22-4e06-baf1-1be4a6fb432a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.851s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1345.367499] env[65385]: WARNING neutronclient.v2_0.client [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1345.368355] env[65385]: WARNING openstack [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1345.368845] env[65385]: WARNING openstack [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1345.477784] env[65385]: WARNING openstack [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1345.478282] env[65385]: WARNING openstack [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1345.553720] env[65385]: WARNING neutronclient.v2_0.client [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1345.554445] env[65385]: WARNING openstack [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1345.554911] env[65385]: WARNING openstack [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1345.636492] env[65385]: DEBUG nova.network.neutron [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Updating instance_info_cache with network_info: [{"id": "078edaec-5c64-4393-b4b9-0e9537ee060c", "address": "fa:16:3e:62:87:51", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap078edaec-5c", "ovs_interfaceid": "078edaec-5c64-4393-b4b9-0e9537ee060c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1345.811961] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1345.812233] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1345.812439] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1345.812610] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1345.812771] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1345.815197] env[65385]: INFO nova.compute.manager [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Terminating instance [ 1346.139394] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1346.319601] env[65385]: DEBUG nova.compute.manager [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1346.319781] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1346.320708] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77988ab-5463-458b-87d6-617754d0f6ea {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.329197] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1346.329472] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-433dafc8-1e82-480c-be9c-b5fda4b4edcb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.337297] env[65385]: DEBUG oslo_vmware.api [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1346.337297] env[65385]: value = "task-4454788" [ 1346.337297] env[65385]: _type = "Task" [ 1346.337297] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.345843] env[65385]: DEBUG oslo_vmware.api [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454788, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.847820] env[65385]: DEBUG oslo_vmware.api [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454788, 'name': PowerOffVM_Task, 'duration_secs': 0.254771} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.848082] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1346.848243] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1346.848508] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29ae1e39-289f-4035-92b4-517df76a55d7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.917931] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1346.918173] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1346.918455] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleting the datastore file [datastore2] c9f09e05-52c4-44fb-bfa8-85bde23ef8ce {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1346.918767] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cd6947d-5e79-46c3-9f3e-3791f311a1ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.925273] env[65385]: DEBUG oslo_vmware.api [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1346.925273] env[65385]: value = "task-4454790" [ 1346.925273] env[65385]: _type = "Task" [ 1346.925273] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.933610] env[65385]: DEBUG oslo_vmware.api [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454790, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.436010] env[65385]: DEBUG oslo_vmware.api [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454790, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134177} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.436313] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1347.436476] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1347.436650] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1347.436819] env[65385]: INFO nova.compute.manager [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1347.437069] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1347.437632] env[65385]: DEBUG nova.compute.manager [-] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1347.437797] env[65385]: DEBUG nova.network.neutron [-] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1347.437984] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1347.438518] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1347.438774] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1347.513633] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1347.675753] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1347.676085] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24d63294-e8a4-4265-a490-1945b7c992ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.684202] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1347.684202] env[65385]: value = "task-4454791" [ 1347.684202] env[65385]: _type = "Task" [ 1347.684202] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.693220] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.749054] env[65385]: DEBUG nova.compute.manager [req-ccc44298-c038-4c07-9e2d-d44211dfef86 req-fa6484d4-1972-41ca-a47b-81b8dbdc9016 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Received event network-vif-deleted-669739ea-2cff-42c7-a832-8b11965d1cdb {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1347.749054] env[65385]: INFO nova.compute.manager [req-ccc44298-c038-4c07-9e2d-d44211dfef86 req-fa6484d4-1972-41ca-a47b-81b8dbdc9016 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Neutron deleted interface 669739ea-2cff-42c7-a832-8b11965d1cdb; detaching it from the instance and deleting it from the info cache [ 1347.749054] env[65385]: DEBUG nova.network.neutron [req-ccc44298-c038-4c07-9e2d-d44211dfef86 req-fa6484d4-1972-41ca-a47b-81b8dbdc9016 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1348.194725] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454791, 'name': PowerOffVM_Task, 'duration_secs': 0.206251} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.194962] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1348.195806] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96b5d36-f5df-4647-a300-c957787a39f6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.214045] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c820ea9-0ea1-4919-a1f9-3265c4c8ff39 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.228153] env[65385]: DEBUG nova.network.neutron [-] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1348.251114] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1c28971-379e-4146-b18e-ce02f6565354 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.255275] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1348.255943] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a131d8c-86c1-4891-9df1-144bb9f76436 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.263712] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db43374-c05c-4d7e-aec7-f639a0215dee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.275729] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1348.275729] env[65385]: value = "task-4454792" [ 1348.275729] env[65385]: _type = "Task" [ 1348.275729] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.284584] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1348.284810] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1348.285096] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1348.285310] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1348.285508] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1348.292682] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e9bd4ec-a08c-4cb1-b299-62204de3e71a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.294817] env[65385]: DEBUG nova.compute.manager [req-ccc44298-c038-4c07-9e2d-d44211dfef86 req-fa6484d4-1972-41ca-a47b-81b8dbdc9016 service nova] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Detach interface failed, port_id=669739ea-2cff-42c7-a832-8b11965d1cdb, reason: Instance c9f09e05-52c4-44fb-bfa8-85bde23ef8ce could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1348.303271] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1348.303440] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1348.304157] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcf9d12a-5f19-4fb8-b918-5c4b1dab266d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.310068] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1348.310068] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e867b2-c266-0ecf-3f0b-7c47c5dcdbf5" [ 1348.310068] env[65385]: _type = "Task" [ 1348.310068] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.318022] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e867b2-c266-0ecf-3f0b-7c47c5dcdbf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.732666] env[65385]: INFO nova.compute.manager [-] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Took 1.29 seconds to deallocate network for instance. [ 1348.822082] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52e867b2-c266-0ecf-3f0b-7c47c5dcdbf5, 'name': SearchDatastore_Task, 'duration_secs': 0.009274} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.822937] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-795a390e-ac6d-4c8d-8cab-0c1b08cdc4eb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.830192] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1348.830192] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5272ac0e-08f0-d77b-9e34-73e038981691" [ 1348.830192] env[65385]: _type = "Task" [ 1348.830192] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.839867] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5272ac0e-08f0-d77b-9e34-73e038981691, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.239357] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1349.239666] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1349.239919] env[65385]: DEBUG nova.objects.instance [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'resources' on Instance uuid c9f09e05-52c4-44fb-bfa8-85bde23ef8ce {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1349.341309] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5272ac0e-08f0-d77b-9e34-73e038981691, 'name': SearchDatastore_Task, 'duration_secs': 0.010798} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.341532] env[65385]: DEBUG oslo_concurrency.lockutils [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1349.341762] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. {{(pid=65385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1349.342047] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6eb714ae-5549-43dd-b9f9-1479df31e503 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.349807] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1349.349807] env[65385]: value = "task-4454793" [ 1349.349807] env[65385]: _type = "Task" [ 1349.349807] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.358189] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454793, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.826752] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1caefed9-4c35-454d-81dc-a88b0a884a60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.834637] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81ae314-fd9d-42fe-9841-67f578f3bd5d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.868046] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87498dbd-e71b-4c0b-8966-be3db8778e12 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.875489] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454793, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448949} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.877538] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. [ 1349.878337] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9d5efd-b1d9-474a-b323-0f2ecf052172 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.881418] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db71211-adda-499b-acf4-b59cf5637873 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.907606] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1349.914852] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f30e3111-7de7-4579-9911-5ff7a865d7d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.928119] env[65385]: DEBUG nova.compute.provider_tree [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1349.934654] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1349.934654] env[65385]: value = "task-4454794" [ 1349.934654] env[65385]: _type = "Task" [ 1349.934654] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.943796] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454794, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.431524] env[65385]: DEBUG nova.scheduler.client.report [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1350.445153] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454794, 'name': ReconfigVM_Task, 'duration_secs': 0.328988} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1350.446341] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Reconfigured VM instance instance-0000007c to attach disk [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1350.446935] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b2c78f-2460-4f67-9182-fb2ee62d677a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.473488] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b7f62ab-da2d-4e34-b1fb-cdb92d755586 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.489740] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1350.489740] env[65385]: value = "task-4454795" [ 1350.489740] env[65385]: _type = "Task" [ 1350.489740] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1350.498817] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454795, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.937327] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1350.960777] env[65385]: INFO nova.scheduler.client.report [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted allocations for instance c9f09e05-52c4-44fb-bfa8-85bde23ef8ce [ 1351.000062] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454795, 'name': ReconfigVM_Task, 'duration_secs': 0.166742} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1351.000437] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1351.000706] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9d17161-3099-4a39-ac7b-83a56fc0186a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.007973] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1351.007973] env[65385]: value = "task-4454796" [ 1351.007973] env[65385]: _type = "Task" [ 1351.007973] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1351.016961] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454796, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.468476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-66c01710-9cce-41e1-b0f9-ebf5974d590d tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "c9f09e05-52c4-44fb-bfa8-85bde23ef8ce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.656s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1351.517983] env[65385]: DEBUG oslo_vmware.api [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454796, 'name': PowerOnVM_Task, 'duration_secs': 0.386918} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1351.518240] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1351.520948] env[65385]: DEBUG nova.compute.manager [None req-98bdc226-a395-4253-ae4c-cb1ba18811b9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1351.521767] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9889aa47-844f-4388-aa77-a5fd1b5aebf8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.901870] env[65385]: INFO nova.compute.manager [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Unrescuing [ 1352.901870] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1352.901870] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1352.901870] env[65385]: DEBUG nova.network.neutron [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1353.405885] env[65385]: WARNING neutronclient.v2_0.client [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1353.406034] env[65385]: WARNING openstack [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1353.406286] env[65385]: WARNING openstack [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1353.533205] env[65385]: WARNING openstack [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1353.533615] env[65385]: WARNING openstack [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1353.599484] env[65385]: WARNING neutronclient.v2_0.client [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1353.600146] env[65385]: WARNING openstack [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1353.600559] env[65385]: WARNING openstack [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1353.685421] env[65385]: DEBUG nova.network.neutron [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Updating instance_info_cache with network_info: [{"id": "078edaec-5c64-4393-b4b9-0e9537ee060c", "address": "fa:16:3e:62:87:51", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap078edaec-5c", "ovs_interfaceid": "078edaec-5c64-4393-b4b9-0e9537ee060c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1354.189029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-02e65e47-7971-4e08-acc9-eead1828d925" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1354.189391] env[65385]: DEBUG nova.objects.instance [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lazy-loading 'flavor' on Instance uuid 02e65e47-7971-4e08-acc9-eead1828d925 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1354.694961] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6b5d92-d89d-4f3a-9243-42f3bcb7f224 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.716548] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1354.716879] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0cf9bed5-9480-4324-bec5-5e1a04bbc1d4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.724377] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1354.724377] env[65385]: value = "task-4454798" [ 1354.724377] env[65385]: _type = "Task" [ 1354.724377] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.734163] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454798, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.235232] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454798, 'name': PowerOffVM_Task, 'duration_secs': 0.252185} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.235670] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1355.240676] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Reconfiguring VM instance instance-0000007c to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1355.240959] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92ce37e1-19f7-488b-8080-faf492c21b4b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.259970] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1355.259970] env[65385]: value = "task-4454799" [ 1355.259970] env[65385]: _type = "Task" [ 1355.259970] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.269375] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454799, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.501436] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1355.501647] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1355.501850] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1355.502031] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1355.502196] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1355.504535] env[65385]: INFO nova.compute.manager [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Terminating instance [ 1355.771242] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454799, 'name': ReconfigVM_Task, 'duration_secs': 0.221884} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.771449] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Reconfigured VM instance instance-0000007c to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1355.771620] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1355.771874] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6076290-9539-4661-806a-f539a141331a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.779436] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1355.779436] env[65385]: value = "task-4454800" [ 1355.779436] env[65385]: _type = "Task" [ 1355.779436] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.791366] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454800, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.008602] env[65385]: DEBUG nova.compute.manager [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1356.008820] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1356.009716] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29dc1c0d-74fb-422f-b758-d9cde234432c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.017889] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1356.018153] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6dc81813-d637-43bd-994c-cb4a9de6678f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.026064] env[65385]: DEBUG oslo_vmware.api [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1356.026064] env[65385]: value = "task-4454801" [ 1356.026064] env[65385]: _type = "Task" [ 1356.026064] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.034526] env[65385]: DEBUG oslo_vmware.api [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454801, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.289651] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454800, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.537122] env[65385]: DEBUG oslo_vmware.api [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454801, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.792023] env[65385]: DEBUG oslo_vmware.api [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454800, 'name': PowerOnVM_Task, 'duration_secs': 0.980989} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.792370] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1356.792626] env[65385]: DEBUG nova.compute.manager [None req-4052027b-5fb8-404b-9129-425c70c0977d tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1356.793470] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d923a19b-cc0f-418a-a515-9b391392bd17 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.040884] env[65385]: DEBUG oslo_vmware.api [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454801, 'name': PowerOffVM_Task, 'duration_secs': 0.578502} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.041384] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1357.041659] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1357.042047] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21f583db-6faa-4b3e-81c8-84872878e54f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.111265] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1357.111495] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1357.111671] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleting the datastore file [datastore2] b4bca2c8-fc8f-471a-899b-aac648fee9fe {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1357.111967] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b73f5aa3-8eab-451b-996f-8b021afa0212 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.119710] env[65385]: DEBUG oslo_vmware.api [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for the task: (returnval){ [ 1357.119710] env[65385]: value = "task-4454803" [ 1357.119710] env[65385]: _type = "Task" [ 1357.119710] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.129235] env[65385]: DEBUG oslo_vmware.api [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.630679] env[65385]: DEBUG oslo_vmware.api [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Task: {'id': task-4454803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140592} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.630949] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1357.631367] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1357.631367] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1357.631483] env[65385]: INFO nova.compute.manager [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1357.631768] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1357.632018] env[65385]: DEBUG nova.compute.manager [-] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1357.632125] env[65385]: DEBUG nova.network.neutron [-] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1357.632369] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1357.632881] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1357.633143] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1357.671600] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1357.956932] env[65385]: DEBUG nova.compute.manager [req-66614fe2-3fc8-423e-9708-0c0e5e8e52f3 req-4e25a4ab-7d32-436c-8546-2afb3827eb34 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Received event network-vif-deleted-c11bab1f-7eb9-434c-8436-c2319951b01f {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1357.957550] env[65385]: INFO nova.compute.manager [req-66614fe2-3fc8-423e-9708-0c0e5e8e52f3 req-4e25a4ab-7d32-436c-8546-2afb3827eb34 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Neutron deleted interface c11bab1f-7eb9-434c-8436-c2319951b01f; detaching it from the instance and deleting it from the info cache [ 1357.957764] env[65385]: DEBUG nova.network.neutron [req-66614fe2-3fc8-423e-9708-0c0e5e8e52f3 req-4e25a4ab-7d32-436c-8546-2afb3827eb34 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1358.430169] env[65385]: DEBUG nova.network.neutron [-] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1358.460661] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2248dfc-a55a-4f67-916d-7b601ab96c8c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.472956] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0ac950-f476-43a8-acab-6c6116e80263 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.502906] env[65385]: DEBUG nova.compute.manager [req-66614fe2-3fc8-423e-9708-0c0e5e8e52f3 req-4e25a4ab-7d32-436c-8546-2afb3827eb34 service nova] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Detach interface failed, port_id=c11bab1f-7eb9-434c-8436-c2319951b01f, reason: Instance b4bca2c8-fc8f-471a-899b-aac648fee9fe could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1358.843792] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1358.932772] env[65385]: INFO nova.compute.manager [-] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Took 1.30 seconds to deallocate network for instance. [ 1359.442135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1359.442545] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1359.442786] env[65385]: DEBUG nova.objects.instance [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lazy-loading 'resources' on Instance uuid b4bca2c8-fc8f-471a-899b-aac648fee9fe {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1359.877350] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1359.877718] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1360.021140] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce87afde-ef18-459e-b648-1e939cbc6250 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.029722] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ce1425-30fe-4119-8f85-6d695e4cb3b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.060540] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d155a742-d5c3-4fef-8f69-391b47e7412a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.069047] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3d3775-9185-43f4-9c7a-fa082cb5da6b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.082690] env[65385]: DEBUG nova.compute.provider_tree [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1360.380376] env[65385]: DEBUG nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1360.586074] env[65385]: DEBUG nova.scheduler.client.report [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1360.904455] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1361.092786] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1361.096372] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.192s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1361.098188] env[65385]: INFO nova.compute.claims [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1361.114969] env[65385]: INFO nova.scheduler.client.report [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Deleted allocations for instance b4bca2c8-fc8f-471a-899b-aac648fee9fe [ 1361.624130] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4c77a809-3b92-424a-b45c-552ad43f5e48 tempest-ServerActionsTestOtherA-753894539 tempest-ServerActionsTestOtherA-753894539-project-member] Lock "b4bca2c8-fc8f-471a-899b-aac648fee9fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.122s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1361.838450] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1361.844348] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1361.844561] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1362.180125] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a94d263-0541-42ed-8218-7e72d561ca66 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.189065] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0505fc22-b291-4317-aa28-bf9d4f27980b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.220130] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1615b11-2f52-4656-8831-7c48ce977e76 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.229693] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ed00c5-7e6c-434c-afb3-21faf204d4dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.245944] env[65385]: DEBUG nova.compute.provider_tree [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1362.347919] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1362.749937] env[65385]: DEBUG nova.scheduler.client.report [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1363.254950] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.159s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1363.255561] env[65385]: DEBUG nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1363.258456] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.911s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1363.258559] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1363.258876] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1363.259801] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29fd3d1-f896-45bd-8d7a-767b1d7e34ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.270939] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3607a93f-b8ea-4c2e-9523-e60fe0c160c3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.289684] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce10f3d5-9c39-4cf7-85f9-52cef4c7338c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.299570] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd9d5b7-845f-4879-bf15-7d99a5db32a8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.333011] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179676MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1363.333208] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1363.333511] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1363.761454] env[65385]: DEBUG nova.compute.utils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1363.762904] env[65385]: DEBUG nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1363.763689] env[65385]: DEBUG nova.network.neutron [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1363.764120] env[65385]: WARNING neutronclient.v2_0.client [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1363.764472] env[65385]: WARNING neutronclient.v2_0.client [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1363.765158] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1363.765571] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1363.817442] env[65385]: DEBUG nova.policy [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07886144152846b487b5d106416ee1bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fc382ad43724fd3ae1b9b62c1dd8131', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1364.124490] env[65385]: DEBUG nova.network.neutron [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Successfully created port: c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1364.273426] env[65385]: DEBUG nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1364.361570] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 6926f1bc-3c40-486f-b3e9-7e3731d57e4e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1364.361721] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 02e65e47-7971-4e08-acc9-eead1828d925 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1364.361839] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance ad85aa5c-68a6-41a3-88d7-c27bbef81c6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1364.361950] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 9ee16fed-c625-4da3-bb83-2b4b2416b99f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1364.362140] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1364.362281] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '2', 'num_task_None': '3', 'num_os_type_None': '4', 'num_proj_df347ee234f142899014703c70ee42ed': '1', 'io_workload': '1', 'num_proj_2fc382ad43724fd3ae1b9b62c1dd8131': '3', 'num_vm_rescued': '1', 'num_vm_building': '1', 'num_task_networking': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1364.426949] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5665098f-c690-42f2-99ee-bc513591c20d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.435200] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2d8cf8-b562-4ae8-849e-53e547a35a6a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.466867] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e17109-2e31-4b94-835a-eb9bf735ebf7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.475053] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfe2d1d-5a40-425e-b7ce-4f947fd6620f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.489626] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1364.993117] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1365.283757] env[65385]: DEBUG nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1365.318970] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1365.319341] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1365.319570] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1365.319898] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1365.320247] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1365.320503] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1365.321026] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1365.321266] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1365.321541] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1365.321821] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1365.322222] env[65385]: DEBUG nova.virt.hardware [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1365.323684] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4476174-c083-4ade-8b03-32fdd1e356bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.336290] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e728f6d6-c355-43cc-bbaa-69e5b2332c5f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.498433] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1365.498636] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.165s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1365.583114] env[65385]: DEBUG nova.compute.manager [req-d42bca3f-8da5-4886-9a17-ae5f0bfd55db req-54d4ce0e-f98e-4e2a-a2fd-327b80055120 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Received event network-vif-plugged-c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1365.583346] env[65385]: DEBUG oslo_concurrency.lockutils [req-d42bca3f-8da5-4886-9a17-ae5f0bfd55db req-54d4ce0e-f98e-4e2a-a2fd-327b80055120 service nova] Acquiring lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1365.584104] env[65385]: DEBUG oslo_concurrency.lockutils [req-d42bca3f-8da5-4886-9a17-ae5f0bfd55db req-54d4ce0e-f98e-4e2a-a2fd-327b80055120 service nova] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1365.584104] env[65385]: DEBUG oslo_concurrency.lockutils [req-d42bca3f-8da5-4886-9a17-ae5f0bfd55db req-54d4ce0e-f98e-4e2a-a2fd-327b80055120 service nova] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1365.584104] env[65385]: DEBUG nova.compute.manager [req-d42bca3f-8da5-4886-9a17-ae5f0bfd55db req-54d4ce0e-f98e-4e2a-a2fd-327b80055120 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] No waiting events found dispatching network-vif-plugged-c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1365.584104] env[65385]: WARNING nova.compute.manager [req-d42bca3f-8da5-4886-9a17-ae5f0bfd55db req-54d4ce0e-f98e-4e2a-a2fd-327b80055120 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Received unexpected event network-vif-plugged-c6668bfd-25c6-4589-b8ee-bbbc79181617 for instance with vm_state building and task_state spawning. [ 1365.724747] env[65385]: DEBUG nova.network.neutron [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Successfully updated port: c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1366.177524] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1366.177825] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1366.228286] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1366.228517] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1366.228638] env[65385]: DEBUG nova.network.neutron [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1366.681246] env[65385]: INFO nova.compute.manager [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Detaching volume 0e1ba855-b903-4512-8811-25211c190139 [ 1366.715405] env[65385]: INFO nova.virt.block_device [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Attempting to driver detach volume 0e1ba855-b903-4512-8811-25211c190139 from mountpoint /dev/sdb [ 1366.715689] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1366.715880] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871218', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'name': 'volume-0e1ba855-b903-4512-8811-25211c190139', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6926f1bc-3c40-486f-b3e9-7e3731d57e4e', 'attached_at': '', 'detached_at': '', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'serial': '0e1ba855-b903-4512-8811-25211c190139'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1366.718339] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3cc142-616d-4129-9fd2-52a42f6d76d1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.740691] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1366.740779] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1366.748525] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323b7659-d490-414b-a936-8fe5faab7f1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.757579] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3674f91d-659d-4d76-ae6b-7761d1ade3ed {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.780320] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442f7f73-d697-43b3-b80d-ce6e98392787 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.785381] env[65385]: DEBUG nova.network.neutron [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1366.801804] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The volume has not been displaced from its original location: [datastore2] volume-0e1ba855-b903-4512-8811-25211c190139/volume-0e1ba855-b903-4512-8811-25211c190139.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1366.807208] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Reconfiguring VM instance instance-00000079 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1366.807612] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cff7fb4a-0bf1-4669-970e-1682145560b4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.821697] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1366.822085] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1366.835535] env[65385]: DEBUG oslo_vmware.api [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1366.835535] env[65385]: value = "task-4454804" [ 1366.835535] env[65385]: _type = "Task" [ 1366.835535] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.844953] env[65385]: DEBUG oslo_vmware.api [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454804, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.910399] env[65385]: WARNING neutronclient.v2_0.client [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1366.911133] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1366.911489] env[65385]: WARNING openstack [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1367.007357] env[65385]: DEBUG nova.network.neutron [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1367.346322] env[65385]: DEBUG oslo_vmware.api [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454804, 'name': ReconfigVM_Task, 'duration_secs': 0.291744} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.346644] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Reconfigured VM instance instance-00000079 to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1367.351412] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8b43cb6-0461-49d2-809b-299fdc408c95 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.368750] env[65385]: DEBUG oslo_vmware.api [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1367.368750] env[65385]: value = "task-4454805" [ 1367.368750] env[65385]: _type = "Task" [ 1367.368750] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.378608] env[65385]: DEBUG oslo_vmware.api [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454805, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.498598] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1367.498808] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1367.498954] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1367.499108] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1367.510154] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1367.510742] env[65385]: DEBUG nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Instance network_info: |[{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1367.511357] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:f5:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1559ce49-7345-443f-bf02-4bfeb88356ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6668bfd-25c6-4589-b8ee-bbbc79181617', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1367.519034] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1367.519238] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1367.519465] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea00373d-cb3d-4d27-bc32-ef056062b9d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.540337] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1367.540337] env[65385]: value = "task-4454806" [ 1367.540337] env[65385]: _type = "Task" [ 1367.540337] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.549084] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454806, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.612458] env[65385]: DEBUG nova.compute.manager [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Received event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1367.612660] env[65385]: DEBUG nova.compute.manager [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing instance network info cache due to event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1367.612916] env[65385]: DEBUG oslo_concurrency.lockutils [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Acquiring lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1367.613103] env[65385]: DEBUG oslo_concurrency.lockutils [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Acquired lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1367.613204] env[65385]: DEBUG nova.network.neutron [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1367.845076] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1367.879669] env[65385]: DEBUG oslo_vmware.api [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454805, 'name': ReconfigVM_Task, 'duration_secs': 0.148386} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.879967] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871218', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'name': 'volume-0e1ba855-b903-4512-8811-25211c190139', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6926f1bc-3c40-486f-b3e9-7e3731d57e4e', 'attached_at': '', 'detached_at': '', 'volume_id': '0e1ba855-b903-4512-8811-25211c190139', 'serial': '0e1ba855-b903-4512-8811-25211c190139'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1368.051843] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454806, 'name': CreateVM_Task, 'duration_secs': 0.323553} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.051843] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1368.052107] env[65385]: WARNING neutronclient.v2_0.client [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1368.052476] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1368.052602] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1368.052952] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1368.053702] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80f76932-7c3b-4a2f-a83f-e680e2a31ca0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.058798] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1368.058798] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52747989-70b5-af82-e385-947a66b59265" [ 1368.058798] env[65385]: _type = "Task" [ 1368.058798] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1368.067666] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52747989-70b5-af82-e385-947a66b59265, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.115635] env[65385]: WARNING neutronclient.v2_0.client [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1368.116422] env[65385]: WARNING openstack [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1368.116962] env[65385]: WARNING openstack [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1368.223917] env[65385]: WARNING openstack [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1368.224341] env[65385]: WARNING openstack [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1368.285902] env[65385]: WARNING neutronclient.v2_0.client [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1368.286658] env[65385]: WARNING openstack [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1368.287236] env[65385]: WARNING openstack [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1368.372386] env[65385]: DEBUG nova.network.neutron [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updated VIF entry in instance network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1368.372749] env[65385]: DEBUG nova.network.neutron [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1368.420876] env[65385]: DEBUG nova.objects.instance [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1368.571485] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52747989-70b5-af82-e385-947a66b59265, 'name': SearchDatastore_Task, 'duration_secs': 0.009884} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.571764] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1368.571988] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1368.572234] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1368.572372] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1368.572539] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1368.572791] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf605a7e-d650-4f6d-825b-74265d87b999 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.581748] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1368.581909] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1368.582604] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5412f2c5-f687-488f-9e79-3ceda2930830 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.588588] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1368.588588] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52077f55-0070-c127-15fb-74b1e0004793" [ 1368.588588] env[65385]: _type = "Task" [ 1368.588588] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1368.596028] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52077f55-0070-c127-15fb-74b1e0004793, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.875415] env[65385]: DEBUG oslo_concurrency.lockutils [req-a2ba8266-740b-4cca-baaf-9f3e7cd49966 req-cba98804-ab5f-4c58-b5e6-f73b7c443a86 service nova] Releasing lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1369.099674] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52077f55-0070-c127-15fb-74b1e0004793, 'name': SearchDatastore_Task, 'duration_secs': 0.018118} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1369.100430] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28aa4db0-4811-40a1-91e8-500c708005ef {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.106247] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1369.106247] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b0e9cc-919f-3b30-dc63-695deb7712ee" [ 1369.106247] env[65385]: _type = "Task" [ 1369.106247] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1369.115166] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b0e9cc-919f-3b30-dc63-695deb7712ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1369.382360] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1369.427723] env[65385]: DEBUG oslo_concurrency.lockutils [None req-c6fc9b53-2604-492f-8fdc-283a6ad0f972 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.250s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1369.429012] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.047s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1369.429632] env[65385]: DEBUG nova.compute.manager [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1369.430288] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c555b6b9-5aa1-405b-b4ce-9240eaf418e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.437599] env[65385]: DEBUG nova.compute.manager [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1369.438249] env[65385]: DEBUG nova.objects.instance [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1369.617369] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52b0e9cc-919f-3b30-dc63-695deb7712ee, 'name': SearchDatastore_Task, 'duration_secs': 0.01021} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1369.617590] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1369.617848] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/9ee16fed-c625-4da3-bb83-2b4b2416b99f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1369.618124] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ae519ce-c556-40eb-b4ef-50542a778926 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.624885] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1369.624885] env[65385]: value = "task-4454807" [ 1369.624885] env[65385]: _type = "Task" [ 1369.624885] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1369.632971] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454807, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.135590] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454807, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44048} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.135873] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/9ee16fed-c625-4da3-bb83-2b4b2416b99f.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1370.136075] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1370.136332] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c90b86a9-657c-47a4-a46a-5642d9672342 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.143832] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1370.143832] env[65385]: value = "task-4454808" [ 1370.143832] env[65385]: _type = "Task" [ 1370.143832] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.152521] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.446324] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1370.446746] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43cf2df6-49b0-41ce-ab26-28ca18f7acba {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.453868] env[65385]: DEBUG oslo_vmware.api [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1370.453868] env[65385]: value = "task-4454809" [ 1370.453868] env[65385]: _type = "Task" [ 1370.453868] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.462643] env[65385]: DEBUG oslo_vmware.api [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454809, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.653970] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061566} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.654270] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1370.655143] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b52ff3-6277-46af-8ad5-c05e8c36f386 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.678648] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfiguring VM instance instance-0000007e to attach disk [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/9ee16fed-c625-4da3-bb83-2b4b2416b99f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1370.678968] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3802747-fa9e-4699-8e0f-13b3307e5cbd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.699736] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1370.699736] env[65385]: value = "task-4454810" [ 1370.699736] env[65385]: _type = "Task" [ 1370.699736] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.709038] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454810, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.966237] env[65385]: DEBUG oslo_vmware.api [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454809, 'name': PowerOffVM_Task, 'duration_secs': 0.209485} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.966601] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1370.966866] env[65385]: DEBUG nova.compute.manager [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1370.967874] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23c80b9-2b96-476c-ac37-c49401f9ee65 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.210562] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454810, 'name': ReconfigVM_Task, 'duration_secs': 0.29903} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.210562] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfigured VM instance instance-0000007e to attach disk [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/9ee16fed-c625-4da3-bb83-2b4b2416b99f.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1371.211210] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab7d09fb-45e6-4f3a-a34b-69a772c138d9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.218010] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1371.218010] env[65385]: value = "task-4454811" [ 1371.218010] env[65385]: _type = "Task" [ 1371.218010] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1371.227715] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454811, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.480475] env[65385]: DEBUG oslo_concurrency.lockutils [None req-d4ccfd79-2e83-4643-8ee3-82b184e62e73 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.051s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1371.728643] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454811, 'name': Rename_Task, 'duration_secs': 0.161343} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.728914] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1371.729179] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1fd546c-80bd-42e6-b17c-600b64d6163c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.736799] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1371.736799] env[65385]: value = "task-4454812" [ 1371.736799] env[65385]: _type = "Task" [ 1371.736799] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1371.744888] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454812, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.808103] env[65385]: DEBUG nova.objects.instance [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1371.839072] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1372.248606] env[65385]: DEBUG oslo_vmware.api [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454812, 'name': PowerOnVM_Task, 'duration_secs': 0.486034} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1372.249030] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1372.249120] env[65385]: INFO nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Took 6.97 seconds to spawn the instance on the hypervisor. [ 1372.249243] env[65385]: DEBUG nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1372.250034] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd0fa05-9842-498b-ab99-d33fe5e83309 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.313267] env[65385]: DEBUG oslo_concurrency.lockutils [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1372.313461] env[65385]: DEBUG oslo_concurrency.lockutils [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1372.313603] env[65385]: DEBUG nova.network.neutron [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1372.313771] env[65385]: DEBUG nova.objects.instance [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'info_cache' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1372.767598] env[65385]: INFO nova.compute.manager [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Took 11.88 seconds to build instance. [ 1372.816791] env[65385]: DEBUG nova.objects.base [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Object Instance<6926f1bc-3c40-486f-b3e9-7e3731d57e4e> lazy-loaded attributes: flavor,info_cache {{(pid=65385) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1373.101489] env[65385]: DEBUG nova.compute.manager [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Received event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1373.101709] env[65385]: DEBUG nova.compute.manager [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing instance network info cache due to event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1373.101915] env[65385]: DEBUG oslo_concurrency.lockutils [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Acquiring lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1373.102059] env[65385]: DEBUG oslo_concurrency.lockutils [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Acquired lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1373.102219] env[65385]: DEBUG nova.network.neutron [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1373.270266] env[65385]: DEBUG oslo_concurrency.lockutils [None req-602bc258-592f-413f-a6cd-7d845938a586 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.392s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1373.320327] env[65385]: WARNING neutronclient.v2_0.client [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1373.320775] env[65385]: WARNING openstack [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.321147] env[65385]: WARNING openstack [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.429804] env[65385]: WARNING openstack [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.430331] env[65385]: WARNING openstack [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.495382] env[65385]: WARNING neutronclient.v2_0.client [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1373.496221] env[65385]: WARNING openstack [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.496654] env[65385]: WARNING openstack [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.577856] env[65385]: DEBUG nova.network.neutron [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating instance_info_cache with network_info: [{"id": "59e7b718-2455-470d-85a3-3a0d0849ea44", "address": "fa:16:3e:7e:65:d9", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59e7b718-24", "ovs_interfaceid": "59e7b718-2455-470d-85a3-3a0d0849ea44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1373.604758] env[65385]: WARNING neutronclient.v2_0.client [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1373.605468] env[65385]: WARNING openstack [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.605858] env[65385]: WARNING openstack [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.726924] env[65385]: WARNING openstack [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.727336] env[65385]: WARNING openstack [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.791843] env[65385]: WARNING neutronclient.v2_0.client [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1373.792596] env[65385]: WARNING openstack [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.793274] env[65385]: WARNING openstack [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.874780] env[65385]: DEBUG nova.network.neutron [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updated VIF entry in instance network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1373.875153] env[65385]: DEBUG nova.network.neutron [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1374.079777] env[65385]: DEBUG oslo_concurrency.lockutils [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "refresh_cache-6926f1bc-3c40-486f-b3e9-7e3731d57e4e" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1374.378378] env[65385]: DEBUG oslo_concurrency.lockutils [req-2eac2a25-acf1-4913-a23a-049d789a2d83 req-b4ca42e8-fcb8-4cc8-952d-3ec9184dc102 service nova] Releasing lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1375.086446] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1375.086803] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43315c8b-cd89-47cc-b29e-2dd0d90db561 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.095473] env[65385]: DEBUG oslo_vmware.api [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1375.095473] env[65385]: value = "task-4454813" [ 1375.095473] env[65385]: _type = "Task" [ 1375.095473] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1375.104251] env[65385]: DEBUG oslo_vmware.api [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1375.607306] env[65385]: DEBUG oslo_vmware.api [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454813, 'name': PowerOnVM_Task, 'duration_secs': 0.385074} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1375.607688] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1375.607835] env[65385]: DEBUG nova.compute.manager [None req-74c0aa0d-ba12-4ec7-a2ca-0a50316d72c4 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1375.608586] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055d7b2e-b092-4d60-bf4c-1dfaea46c429 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.183043] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1409.183397] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1409.687460] env[65385]: DEBUG nova.compute.utils [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1410.190974] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1411.260206] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1411.260613] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1411.260680] env[65385]: INFO nova.compute.manager [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Attaching volume 4d9484bb-eb7b-4915-ab78-a5d252589350 to /dev/sdb [ 1411.291666] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60548ee4-4ad3-485a-8266-5bcbe6d0dcd8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.299973] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a49e6c5-44fa-4b2d-abd2-07275172ea28 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.314028] env[65385]: DEBUG nova.virt.block_device [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating existing volume attachment record: 2b42255c-681e-42cc-b31f-e1867b9f25aa {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1412.238023] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1412.238311] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1412.238521] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1412.238703] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1412.238860] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1412.242494] env[65385]: INFO nova.compute.manager [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Terminating instance [ 1412.747844] env[65385]: DEBUG nova.compute.manager [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1412.748226] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1412.749452] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7250ba1f-5b53-4b51-ad81-2d062c241808 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.757740] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1412.757996] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a82e035-ceb5-45d2-bff4-04b8bdfb6c7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.764691] env[65385]: DEBUG oslo_vmware.api [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1412.764691] env[65385]: value = "task-4454817" [ 1412.764691] env[65385]: _type = "Task" [ 1412.764691] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1412.773500] env[65385]: DEBUG oslo_vmware.api [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1413.275415] env[65385]: DEBUG oslo_vmware.api [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454817, 'name': PowerOffVM_Task, 'duration_secs': 0.191936} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1413.275713] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1413.275834] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1413.276111] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b43bd094-33f7-47cf-a44c-e140b3aaad4d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.350467] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1413.350684] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1413.350847] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Deleting the datastore file [datastore1] 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1413.351140] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0993dc43-515e-4a47-a1fd-5df15ebbcc20 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.357972] env[65385]: DEBUG oslo_vmware.api [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1413.357972] env[65385]: value = "task-4454819" [ 1413.357972] env[65385]: _type = "Task" [ 1413.357972] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1413.366582] env[65385]: DEBUG oslo_vmware.api [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1413.868263] env[65385]: DEBUG oslo_vmware.api [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131196} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1413.868533] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1413.868690] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1413.868867] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1413.869133] env[65385]: INFO nova.compute.manager [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1413.869297] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1413.869486] env[65385]: DEBUG nova.compute.manager [-] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1413.869625] env[65385]: DEBUG nova.network.neutron [-] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1413.869896] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1413.870446] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1413.870715] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1413.907057] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1414.365970] env[65385]: DEBUG nova.compute.manager [req-f2e99d7c-c3c9-4420-9111-d38209133432 req-5b8e91cc-d222-4180-8616-c264c8b63b0e service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Received event network-vif-deleted-59e7b718-2455-470d-85a3-3a0d0849ea44 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1414.366271] env[65385]: INFO nova.compute.manager [req-f2e99d7c-c3c9-4420-9111-d38209133432 req-5b8e91cc-d222-4180-8616-c264c8b63b0e service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Neutron deleted interface 59e7b718-2455-470d-85a3-3a0d0849ea44; detaching it from the instance and deleting it from the info cache [ 1414.366485] env[65385]: DEBUG nova.network.neutron [req-f2e99d7c-c3c9-4420-9111-d38209133432 req-5b8e91cc-d222-4180-8616-c264c8b63b0e service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1414.844467] env[65385]: DEBUG nova.network.neutron [-] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1414.869412] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da1b3183-1bcc-46c1-87cc-01f3d126c664 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.880920] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8841bed6-7ef7-4aa0-875d-78c23f594e94 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.908530] env[65385]: DEBUG nova.compute.manager [req-f2e99d7c-c3c9-4420-9111-d38209133432 req-5b8e91cc-d222-4180-8616-c264c8b63b0e service nova] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Detach interface failed, port_id=59e7b718-2455-470d-85a3-3a0d0849ea44, reason: Instance 6926f1bc-3c40-486f-b3e9-7e3731d57e4e could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1415.348104] env[65385]: INFO nova.compute.manager [-] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Took 1.48 seconds to deallocate network for instance. [ 1415.854784] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1415.855177] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1415.855425] env[65385]: DEBUG nova.objects.instance [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'resources' on Instance uuid 6926f1bc-3c40-486f-b3e9-7e3731d57e4e {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1415.857448] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1415.857654] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871225', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'name': 'volume-4d9484bb-eb7b-4915-ab78-a5d252589350', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9ee16fed-c625-4da3-bb83-2b4b2416b99f', 'attached_at': '', 'detached_at': '', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'serial': '4d9484bb-eb7b-4915-ab78-a5d252589350'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1415.858507] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ac4c42-6d14-4f87-a42b-797edd9f5385 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.876135] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e17f783-3cf8-44da-a847-acf1fef1e5db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.901631] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfiguring VM instance instance-0000007e to attach disk [datastore1] volume-4d9484bb-eb7b-4915-ab78-a5d252589350/volume-4d9484bb-eb7b-4915-ab78-a5d252589350.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1415.901936] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edc56c10-751f-4d84-b8da-d46457b702c8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.921432] env[65385]: DEBUG oslo_vmware.api [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1415.921432] env[65385]: value = "task-4454821" [ 1415.921432] env[65385]: _type = "Task" [ 1415.921432] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1415.930200] env[65385]: DEBUG oslo_vmware.api [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454821, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.432468] env[65385]: DEBUG oslo_vmware.api [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454821, 'name': ReconfigVM_Task, 'duration_secs': 0.334205} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.433634] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfigured VM instance instance-0000007e to attach disk [datastore1] volume-4d9484bb-eb7b-4915-ab78-a5d252589350/volume-4d9484bb-eb7b-4915-ab78-a5d252589350.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1416.438701] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ec2b86-683b-4c8b-96c5-3fdbc2f562bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.441233] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ea04e3e-4f09-4a6b-8245-e36fc6f37640 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.457725] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234b31cf-25a1-4bd8-8103-45dce05690c7 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.460842] env[65385]: DEBUG oslo_vmware.api [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1416.460842] env[65385]: value = "task-4454822" [ 1416.460842] env[65385]: _type = "Task" [ 1416.460842] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.491837] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f32676-c7cf-437d-8052-226020e07bbd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.497700] env[65385]: DEBUG oslo_vmware.api [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454822, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.503233] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ff196e-267d-403d-90a2-313a146bcbf2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.518101] env[65385]: DEBUG nova.compute.provider_tree [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1416.971720] env[65385]: DEBUG oslo_vmware.api [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454822, 'name': ReconfigVM_Task, 'duration_secs': 0.153081} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.972118] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871225', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'name': 'volume-4d9484bb-eb7b-4915-ab78-a5d252589350', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9ee16fed-c625-4da3-bb83-2b4b2416b99f', 'attached_at': '', 'detached_at': '', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'serial': '4d9484bb-eb7b-4915-ab78-a5d252589350'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1417.021334] env[65385]: DEBUG nova.scheduler.client.report [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1417.526500] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1417.547505] env[65385]: INFO nova.scheduler.client.report [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Deleted allocations for instance 6926f1bc-3c40-486f-b3e9-7e3731d57e4e [ 1418.008977] env[65385]: DEBUG nova.objects.instance [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lazy-loading 'flavor' on Instance uuid 9ee16fed-c625-4da3-bb83-2b4b2416b99f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1418.055688] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b21b124c-1c80-466c-b965-0df599d988db tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "6926f1bc-3c40-486f-b3e9-7e3731d57e4e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.817s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1418.514624] env[65385]: DEBUG oslo_concurrency.lockutils [None req-9a593870-4f8a-43f0-8929-3bcd51034bef tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.254s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1418.678718] env[65385]: INFO nova.compute.manager [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Rescuing [ 1418.679029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1418.679187] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1418.679352] env[65385]: DEBUG nova.network.neutron [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1419.181994] env[65385]: WARNING neutronclient.v2_0.client [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1419.182883] env[65385]: WARNING openstack [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1419.183292] env[65385]: WARNING openstack [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1419.286538] env[65385]: WARNING openstack [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1419.286955] env[65385]: WARNING openstack [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1419.346142] env[65385]: WARNING neutronclient.v2_0.client [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1419.346289] env[65385]: WARNING openstack [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1419.346719] env[65385]: WARNING openstack [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1419.428069] env[65385]: DEBUG nova.network.neutron [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1419.930084] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1420.867649] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1420.870624] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1420.870787] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Cleaning up deleted instances with incomplete migration {{(pid=65385) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11954}} [ 1420.871779] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1420.871992] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1421.373802] env[65385]: DEBUG nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Starting instance... {{(pid=65385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1421.881474] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1421.883998] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4ad44d7-9ce0-4a4d-8f08-e3358aabbeda {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.894289] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1421.894289] env[65385]: value = "task-4454823" [ 1421.894289] env[65385]: _type = "Task" [ 1421.894289] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.904520] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454823, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.906018] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1421.906286] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1421.907967] env[65385]: INFO nova.compute.claims [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1422.346978] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1422.347199] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1422.404021] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454823, 'name': PowerOffVM_Task, 'duration_secs': 0.208156} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1422.404300] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1422.405133] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10e71d7-bd85-455c-a4e6-426330a528ad {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.428544] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a687f18-fc8d-4621-8ec8-38878cefe82e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.461634] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1422.461910] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-404778c6-5c47-4c39-ae27-22f8dd5cff40 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.469440] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1422.469440] env[65385]: value = "task-4454824" [ 1422.469440] env[65385]: _type = "Task" [ 1422.469440] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1422.478047] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1422.850126] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1422.983627] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] VM already powered off {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1422.983932] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1422.984087] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1422.984231] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1422.984409] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1422.984869] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1dd438a6-ba79-43f2-aab8-75ff20308430 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.994600] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1422.994783] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1422.995534] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88821471-d23c-4eab-887e-5412a2db887f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.000057] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57739190-1b37-4c90-93d8-53ffe98b72e8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.004668] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1423.004668] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52042b4b-09cc-a2af-6d8b-0506db6d8f2b" [ 1423.004668] env[65385]: _type = "Task" [ 1423.004668] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.010512] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03a656f-18c0-4409-bd10-2b9ae0ac3f07 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.021039] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52042b4b-09cc-a2af-6d8b-0506db6d8f2b, 'name': SearchDatastore_Task, 'duration_secs': 0.010212} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1423.043959] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6739e825-ccf3-4119-b6f2-7d46702b28e6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.046964] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7c8cdc-0852-45f5-bb4b-c0e0f5b45638 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.053721] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1423.053721] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c11aa3-8676-0241-0b48-1b02837c2b7b" [ 1423.053721] env[65385]: _type = "Task" [ 1423.053721] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.059607] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99da31b-0d5e-4f74-a7ef-5c5b91c08edb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.069079] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52c11aa3-8676-0241-0b48-1b02837c2b7b, 'name': SearchDatastore_Task, 'duration_secs': 0.01075} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1423.077927] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1423.078192] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. {{(pid=65385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1423.078696] env[65385]: DEBUG nova.compute.provider_tree [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1423.080047] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83a3a00d-eca7-4a6c-aabe-d87ccb0f3084 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.088787] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1423.088787] env[65385]: value = "task-4454825" [ 1423.088787] env[65385]: _type = "Task" [ 1423.088787] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.097852] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454825, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1423.584145] env[65385]: DEBUG nova.scheduler.client.report [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1423.598958] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454825, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484458} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1423.599892] env[65385]: INFO nova.virt.vmwareapi.ds_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk. [ 1423.600688] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8c3e5a-88df-4763-98fe-45907149fa73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.628773] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfiguring VM instance instance-0000007e to attach disk [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1423.629051] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2abc977-63af-4754-bce1-408fee1906e3 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.648463] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1423.648463] env[65385]: value = "task-4454826" [ 1423.648463] env[65385]: _type = "Task" [ 1423.648463] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.657212] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454826, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.089437] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1424.090027] env[65385]: DEBUG nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Start building networks asynchronously for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1424.092770] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.243s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1424.092952] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1424.093113] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1424.094127] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbd3433-6bed-4cd4-a105-cb6227359383 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.102720] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c8972f-1fb5-41d3-b2a1-c7f1acd724d8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.116668] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3384cbb-71f6-4bed-9e2f-210ce700e1f4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.124300] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafe9db1-dd0a-463e-b529-c138d537398b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.154353] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179382MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1424.154515] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1424.154708] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1424.164799] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454826, 'name': ReconfigVM_Task, 'duration_secs': 0.342662} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.165078] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfigured VM instance instance-0000007e to attach disk [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f/c3e9f50a-a10e-4952-ab03-d986996039fa-rescue.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1424.165887] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a156fe-f0b6-4624-aadb-b088edc85109 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.194168] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67ea7eb4-829b-43c4-b8a3-a7ce34fd3b80 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.209950] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1424.209950] env[65385]: value = "task-4454827" [ 1424.209950] env[65385]: _type = "Task" [ 1424.209950] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.218321] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454827, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.595456] env[65385]: DEBUG nova.compute.utils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1424.596959] env[65385]: DEBUG nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Allocating IP information in the background. {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1424.597165] env[65385]: DEBUG nova.network.neutron [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] allocate_for_instance() {{(pid=65385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1424.597508] env[65385]: WARNING neutronclient.v2_0.client [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1424.597845] env[65385]: WARNING neutronclient.v2_0.client [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1424.598444] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1424.598785] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1424.641896] env[65385]: DEBUG nova.policy [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1394af239e8482d8efd42d6cdc96b5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df347ee234f142899014703c70ee42ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65385) authorize /opt/stack/nova/nova/policy.py:192}} [ 1424.719845] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454827, 'name': ReconfigVM_Task, 'duration_secs': 0.193447} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.720130] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1424.720385] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce2402d0-eee7-42d6-a247-b76c5f548228 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.728865] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1424.728865] env[65385]: value = "task-4454828" [ 1424.728865] env[65385]: _type = "Task" [ 1424.728865] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.741598] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454828, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.964184] env[65385]: DEBUG nova.network.neutron [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Successfully created port: f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1425.106723] env[65385]: DEBUG nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Start building block device mappings for instance. {{(pid=65385) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1425.240396] env[65385]: DEBUG oslo_vmware.api [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454828, 'name': PowerOnVM_Task, 'duration_secs': 0.424407} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1425.240674] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1425.243592] env[65385]: DEBUG nova.compute.manager [None req-8d466f6e-5369-49c6-814c-e6101f774b9f tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1425.244494] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714d5c46-7c71-4e94-97f7-ea7afdf293c9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.312840] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 02e65e47-7971-4e08-acc9-eead1828d925 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1425.312993] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance ad85aa5c-68a6-41a3-88d7-c27bbef81c6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1425.313113] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 9ee16fed-c625-4da3-bb83-2b4b2416b99f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1425.313219] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 218379e3-49ca-4fdf-8c21-757ae53a1653 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1425.313399] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1425.313542] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '2', 'num_task_None': '2', 'num_os_type_None': '4', 'num_proj_2fc382ad43724fd3ae1b9b62c1dd8131': '3', 'io_workload': '2', 'num_vm_rescued': '1', 'num_task_rescuing': '1', 'num_vm_building': '1', 'num_task_networking': '1', 'num_proj_df347ee234f142899014703c70ee42ed': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1425.381422] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f79038-7c28-4943-90c3-56e91d43ef0b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.390207] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee2a796-fbe5-4222-98f2-d058bb70ddfc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.421099] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9eb4d98-5186-4ff7-8dba-535d23698b04 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.430196] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1243cbf2-d3e3-4da9-86d3-bc1479d53501 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.445101] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1425.948839] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1426.117572] env[65385]: DEBUG nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Start spawning the instance on the hypervisor. {{(pid=65385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1426.147307] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-14T16:44:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-14T16:44:29Z,direct_url=,disk_format='vmdk',id=c3e9f50a-a10e-4952-ab03-d986996039fa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='51b61584e66c48498e8f79f7febc4707',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-14T16:44:30Z,virtual_size=,visibility=), allow threads: False {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1426.147556] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Flavor limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1426.147757] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Image limits 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1426.147955] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Flavor pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1426.148114] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Image pref 0:0:0 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1426.148291] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1426.148505] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1426.148660] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1426.148854] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Got 1 possible topologies {{(pid=65385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1426.149045] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1426.149221] env[65385]: DEBUG nova.virt.hardware [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1426.150158] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e0fccc-4a70-44a9-bffc-5b59bd0944af {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.159138] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9278757b-962d-4ae6-80ce-8b3b1b830e89 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.454067] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1426.454284] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.300s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1426.462050] env[65385]: DEBUG nova.compute.manager [req-57f73605-6110-4496-8527-22f36e343d83 req-107d8ffe-793f-42c7-937d-994faa9fc682 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Received event network-vif-plugged-f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1426.462050] env[65385]: DEBUG oslo_concurrency.lockutils [req-57f73605-6110-4496-8527-22f36e343d83 req-107d8ffe-793f-42c7-937d-994faa9fc682 service nova] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1426.462050] env[65385]: DEBUG oslo_concurrency.lockutils [req-57f73605-6110-4496-8527-22f36e343d83 req-107d8ffe-793f-42c7-937d-994faa9fc682 service nova] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1426.462311] env[65385]: DEBUG oslo_concurrency.lockutils [req-57f73605-6110-4496-8527-22f36e343d83 req-107d8ffe-793f-42c7-937d-994faa9fc682 service nova] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1426.462311] env[65385]: DEBUG nova.compute.manager [req-57f73605-6110-4496-8527-22f36e343d83 req-107d8ffe-793f-42c7-937d-994faa9fc682 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] No waiting events found dispatching network-vif-plugged-f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1426.462386] env[65385]: WARNING nova.compute.manager [req-57f73605-6110-4496-8527-22f36e343d83 req-107d8ffe-793f-42c7-937d-994faa9fc682 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Received unexpected event network-vif-plugged-f60c60da-40d3-44a8-b2bd-0d0464f0e773 for instance with vm_state building and task_state spawning. [ 1426.562388] env[65385]: DEBUG nova.network.neutron [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Successfully updated port: f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1426.919532] env[65385]: INFO nova.compute.manager [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Unrescuing [ 1426.919893] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1426.920056] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquired lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1426.920224] env[65385]: DEBUG nova.network.neutron [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1427.066087] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1427.066292] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1427.066640] env[65385]: DEBUG nova.network.neutron [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Building network info cache for instance {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1427.423682] env[65385]: WARNING neutronclient.v2_0.client [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1427.424619] env[65385]: WARNING openstack [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.425089] env[65385]: WARNING openstack [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.545788] env[65385]: WARNING openstack [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.546220] env[65385]: WARNING openstack [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.569611] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.570055] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.604851] env[65385]: DEBUG nova.network.neutron [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Instance cache missing network info. {{(pid=65385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1427.627903] env[65385]: WARNING neutronclient.v2_0.client [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1427.628598] env[65385]: WARNING openstack [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.628995] env[65385]: WARNING openstack [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.638831] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.639206] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.700567] env[65385]: WARNING neutronclient.v2_0.client [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1427.701232] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.701569] env[65385]: WARNING openstack [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.712120] env[65385]: DEBUG nova.network.neutron [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1427.780726] env[65385]: DEBUG nova.network.neutron [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updating instance_info_cache with network_info: [{"id": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "address": "fa:16:3e:5a:f9:f5", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf60c60da-40", "ovs_interfaceid": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1427.844260] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1427.844492] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1427.844738] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1427.844911] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1427.845107] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1427.845222] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1427.845338] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Cleaning up deleted instances {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11916}} [ 1428.214963] env[65385]: DEBUG oslo_concurrency.lockutils [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Releasing lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1428.215670] env[65385]: DEBUG nova.objects.instance [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lazy-loading 'flavor' on Instance uuid 9ee16fed-c625-4da3-bb83-2b4b2416b99f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1428.283179] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1428.283559] env[65385]: DEBUG nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Instance network_info: |[{"id": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "address": "fa:16:3e:5a:f9:f5", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf60c60da-40", "ovs_interfaceid": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1428.284065] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:f9:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ebd8af-aaf6-4d04-b869-3882e2571ed7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f60c60da-40d3-44a8-b2bd-0d0464f0e773', 'vif_model': 'vmxnet3'}] {{(pid=65385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1428.291488] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1428.291718] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Creating VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1428.291950] env[65385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-acd58229-395e-48f8-8f1b-fa8d98b4e09f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.313899] env[65385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1428.313899] env[65385]: value = "task-4454829" [ 1428.313899] env[65385]: _type = "Task" [ 1428.313899] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.322734] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454829, 'name': CreateVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.355280] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] There are 26 instances to clean {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11925}} [ 1428.355456] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5d2820b9-db76-4572-b0ab-3c4a3ced0fdb] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1428.490807] env[65385]: DEBUG nova.compute.manager [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Received event network-changed-f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1428.490807] env[65385]: DEBUG nova.compute.manager [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Refreshing instance network info cache due to event network-changed-f60c60da-40d3-44a8-b2bd-0d0464f0e773. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1428.490807] env[65385]: DEBUG oslo_concurrency.lockutils [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Acquiring lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1428.490807] env[65385]: DEBUG oslo_concurrency.lockutils [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Acquired lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1428.491220] env[65385]: DEBUG nova.network.neutron [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Refreshing network info cache for port f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1428.721158] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4e1449-c1aa-4505-8040-3b7bdb9d1624 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.745437] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1428.745791] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52f44b83-222c-4965-b483-93838669f969 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.753081] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1428.753081] env[65385]: value = "task-4454830" [ 1428.753081] env[65385]: _type = "Task" [ 1428.753081] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.762454] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.824621] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454829, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.858388] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 2c15094a-1c22-4e06-baf1-1be4a6fb432a] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1428.993233] env[65385]: WARNING neutronclient.v2_0.client [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1428.993908] env[65385]: WARNING openstack [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1428.994274] env[65385]: WARNING openstack [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1429.092432] env[65385]: WARNING openstack [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1429.092839] env[65385]: WARNING openstack [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1429.150789] env[65385]: WARNING neutronclient.v2_0.client [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1429.151467] env[65385]: WARNING openstack [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1429.151808] env[65385]: WARNING openstack [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1429.233229] env[65385]: DEBUG nova.network.neutron [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updated VIF entry in instance network info cache for port f60c60da-40d3-44a8-b2bd-0d0464f0e773. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1429.233605] env[65385]: DEBUG nova.network.neutron [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updating instance_info_cache with network_info: [{"id": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "address": "fa:16:3e:5a:f9:f5", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf60c60da-40", "ovs_interfaceid": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1429.263434] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454830, 'name': PowerOffVM_Task, 'duration_secs': 0.211312} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.263637] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1429.268915] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfiguring VM instance instance-0000007e to detach disk 2002 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1429.269231] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2eabc3ad-31bf-4def-bf49-a940ae0c1472 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.288486] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1429.288486] env[65385]: value = "task-4454831" [ 1429.288486] env[65385]: _type = "Task" [ 1429.288486] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.297145] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454831, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.325456] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454829, 'name': CreateVM_Task} progress is 99%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.362276] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 6926f1bc-3c40-486f-b3e9-7e3731d57e4e] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1429.736635] env[65385]: DEBUG oslo_concurrency.lockutils [req-e420d36f-82a0-4044-a297-320ef8a01af2 req-dbaefd71-977f-44bf-8d9c-80a6fcc2949e service nova] Releasing lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1429.799129] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454831, 'name': ReconfigVM_Task, 'duration_secs': 0.245662} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.799438] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfigured VM instance instance-0000007e to detach disk 2002 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1429.799616] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1429.799871] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54e59ad7-5872-4448-a8ad-e93a72f7b3ec {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.807344] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1429.807344] env[65385]: value = "task-4454832" [ 1429.807344] env[65385]: _type = "Task" [ 1429.807344] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.817467] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.825521] env[65385]: DEBUG oslo_vmware.api [-] Task: {'id': task-4454829, 'name': CreateVM_Task, 'duration_secs': 1.360739} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.825697] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Created VM on the ESX host {{(pid=65385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1429.826187] env[65385]: WARNING neutronclient.v2_0.client [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1429.826547] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1429.826739] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1429.827060] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1429.827314] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b313ef1f-4986-4990-b191-f41ec3346ffd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.832516] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1429.832516] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c9878-ec77-175c-f560-0afdb412ac2f" [ 1429.832516] env[65385]: _type = "Task" [ 1429.832516] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.846112] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c9878-ec77-175c-f560-0afdb412ac2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.865840] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 120aa07c-d0c1-4031-8b61-986a261f2291] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1430.319264] env[65385]: DEBUG oslo_vmware.api [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454832, 'name': PowerOnVM_Task, 'duration_secs': 0.376528} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.319521] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1430.319745] env[65385]: DEBUG nova.compute.manager [None req-4028c83e-87aa-4596-9c35-b9da7e641d16 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1430.320507] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58765fb5-731a-47ff-9f09-7bd100d0d859 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.342162] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]521c9878-ec77-175c-f560-0afdb412ac2f, 'name': SearchDatastore_Task, 'duration_secs': 0.02347} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.342447] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1430.342660] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Processing image c3e9f50a-a10e-4952-ab03-d986996039fa {{(pid=65385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1430.342882] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1430.343026] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1430.343196] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1430.343446] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c144bfeb-b710-4b01-a660-40f026bdd871 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.353105] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1430.353312] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1430.354060] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3eade6f-52b7-4607-a6dc-744051fadf96 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.359784] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1430.359784] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290b01b-afa2-5b06-247b-46247e1cde1c" [ 1430.359784] env[65385]: _type = "Task" [ 1430.359784] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.369320] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: c9f09e05-52c4-44fb-bfa8-85bde23ef8ce] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1430.371041] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290b01b-afa2-5b06-247b-46247e1cde1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.872506] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 447d0bce-8995-4bef-bf1a-58269898a461] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1430.874452] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]5290b01b-afa2-5b06-247b-46247e1cde1c, 'name': SearchDatastore_Task, 'duration_secs': 0.011736} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.875775] env[65385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ab56f0d-5ca2-4af9-b1ae-e8dbf6098483 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.881744] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1430.881744] env[65385]: value = "session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa809e-8893-af9a-95d5-d48f81a0046f" [ 1430.881744] env[65385]: _type = "Task" [ 1430.881744] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.891854] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa809e-8893-af9a-95d5-d48f81a0046f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.376365] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: fa02564b-c1a8-4f8f-876c-806618e37f5b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1431.393197] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': session[529ddb02-0ff9-d501-87e2-4033416e9fc7]52fa809e-8893-af9a-95d5-d48f81a0046f, 'name': SearchDatastore_Task, 'duration_secs': 0.010644} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.393413] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1431.393655] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 218379e3-49ca-4fdf-8c21-757ae53a1653/218379e3-49ca-4fdf-8c21-757ae53a1653.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1431.393923] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aaf07e8c-4957-40b5-8b2c-1a297fff4322 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.401224] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1431.401224] env[65385]: value = "task-4454833" [ 1431.401224] env[65385]: _type = "Task" [ 1431.401224] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.411826] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454833, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.458111] env[65385]: DEBUG nova.compute.manager [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Received event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1431.458266] env[65385]: DEBUG nova.compute.manager [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing instance network info cache due to event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1431.458509] env[65385]: DEBUG oslo_concurrency.lockutils [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Acquiring lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1431.458654] env[65385]: DEBUG oslo_concurrency.lockutils [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Acquired lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1431.458827] env[65385]: DEBUG nova.network.neutron [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1431.879882] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 9de8e3db-35b5-43a8-b677-d815055a0a51] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1431.911535] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454833, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.961587] env[65385]: WARNING neutronclient.v2_0.client [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1431.962316] env[65385]: WARNING openstack [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1431.962652] env[65385]: WARNING openstack [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1432.083809] env[65385]: WARNING openstack [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1432.084517] env[65385]: WARNING openstack [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1432.146915] env[65385]: WARNING neutronclient.v2_0.client [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1432.147616] env[65385]: WARNING openstack [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1432.147964] env[65385]: WARNING openstack [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1432.228655] env[65385]: DEBUG nova.network.neutron [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updated VIF entry in instance network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1432.229072] env[65385]: DEBUG nova.network.neutron [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1432.382977] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: e5cb37e7-0b31-4fd0-b841-ea205f2e6b8f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1432.412195] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454833, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535336} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.412453] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c3e9f50a-a10e-4952-ab03-d986996039fa/c3e9f50a-a10e-4952-ab03-d986996039fa.vmdk to [datastore1] 218379e3-49ca-4fdf-8c21-757ae53a1653/218379e3-49ca-4fdf-8c21-757ae53a1653.vmdk {{(pid=65385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1432.412660] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Extending root virtual disk to 1048576 {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1432.412917] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aeeb0fe8-1ea4-4b9c-b7db-8b1cbb16ffdc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.421107] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1432.421107] env[65385]: value = "task-4454834" [ 1432.421107] env[65385]: _type = "Task" [ 1432.421107] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.429852] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454834, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1432.731736] env[65385]: DEBUG oslo_concurrency.lockutils [req-f99a518d-d039-4c59-ac17-5ccba1a1dca8 req-f0bb4611-4fa1-40d9-8c33-7ab5c011e9f6 service nova] Releasing lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1432.887175] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 8db87973-e093-4da1-8534-62ea6896a57d] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1432.931494] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454834, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.209559} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.931749] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Extended root virtual disk {{(pid=65385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1432.932537] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7565f8-ac19-4958-b831-1c2ec5f5ce56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.955101] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfiguring VM instance instance-0000007f to attach disk [datastore1] 218379e3-49ca-4fdf-8c21-757ae53a1653/218379e3-49ca-4fdf-8c21-757ae53a1653.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1432.955369] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4edd7513-dd1a-4987-8d3b-f65f9a7196ab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.975621] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1432.975621] env[65385]: value = "task-4454835" [ 1432.975621] env[65385]: _type = "Task" [ 1432.975621] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.984900] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454835, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.390829] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 20933973-82ae-498c-b016-3a82bb5dc165] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1433.484865] env[65385]: DEBUG nova.compute.manager [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Received event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1433.485077] env[65385]: DEBUG nova.compute.manager [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing instance network info cache due to event network-changed-c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1433.485294] env[65385]: DEBUG oslo_concurrency.lockutils [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Acquiring lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1433.485464] env[65385]: DEBUG oslo_concurrency.lockutils [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Acquired lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1433.485743] env[65385]: DEBUG nova.network.neutron [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Refreshing network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1433.490882] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454835, 'name': ReconfigVM_Task, 'duration_secs': 0.335549} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1433.491420] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfigured VM instance instance-0000007f to attach disk [datastore1] 218379e3-49ca-4fdf-8c21-757ae53a1653/218379e3-49ca-4fdf-8c21-757ae53a1653.vmdk or device None with type sparse {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1433.492025] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-354e807c-143d-4533-9bc3-52e20d6f8e82 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.500450] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1433.500450] env[65385]: value = "task-4454836" [ 1433.500450] env[65385]: _type = "Task" [ 1433.500450] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1433.512023] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454836, 'name': Rename_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.893860] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 5873d9d9-ff2d-4228-a4ed-b07717ba082e] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1433.992779] env[65385]: WARNING neutronclient.v2_0.client [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1433.993490] env[65385]: WARNING openstack [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1433.993864] env[65385]: WARNING openstack [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1434.011301] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454836, 'name': Rename_Task} progress is 14%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.105635] env[65385]: WARNING openstack [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1434.106085] env[65385]: WARNING openstack [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1434.164554] env[65385]: WARNING neutronclient.v2_0.client [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1434.165276] env[65385]: WARNING openstack [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1434.165615] env[65385]: WARNING openstack [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1434.244012] env[65385]: DEBUG nova.network.neutron [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updated VIF entry in instance network info cache for port c6668bfd-25c6-4589-b8ee-bbbc79181617. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1434.244395] env[65385]: DEBUG nova.network.neutron [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [{"id": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "address": "fa:16:3e:1c:f5:0d", "network": {"id": "2800a63d-2757-4823-88a0-85c7958b0062", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1266686690-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2fc382ad43724fd3ae1b9b62c1dd8131", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1559ce49-7345-443f-bf02-4bfeb88356ef", "external-id": "nsx-vlan-transportzone-670", "segmentation_id": 670, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6668bfd-25", "ovs_interfaceid": "c6668bfd-25c6-4589-b8ee-bbbc79181617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1434.396819] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 54faee8c-6b2d-4864-bf96-26b9dfba583b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1434.511239] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454836, 'name': Rename_Task, 'duration_secs': 0.889069} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.511540] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Powering on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1434.511744] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10ed97dd-5223-496b-b8f4-e71dd13e079e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.518625] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1434.518625] env[65385]: value = "task-4454837" [ 1434.518625] env[65385]: _type = "Task" [ 1434.518625] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1434.526581] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454837, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.747941] env[65385]: DEBUG oslo_concurrency.lockutils [req-8a33b446-3abe-46f1-9755-c824d064967a req-bb4f825c-6951-42f7-aa55-4e7324f14008 service nova] Releasing lock "refresh_cache-9ee16fed-c625-4da3-bb83-2b4b2416b99f" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1434.901434] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 51796586-5084-4656-8a04-e01e7b0847cf] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1435.028526] env[65385]: DEBUG oslo_vmware.api [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454837, 'name': PowerOnVM_Task, 'duration_secs': 0.434781} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1435.028797] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Powered on the VM {{(pid=65385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1435.029038] env[65385]: INFO nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Took 8.91 seconds to spawn the instance on the hypervisor. [ 1435.029269] env[65385]: DEBUG nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Checking state {{(pid=65385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1435.030100] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef004429-8272-4011-9a5e-ccf7d02e2dbe {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.404976] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 62c30ded-c851-477f-b1d4-921268a6ef1c] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1435.547369] env[65385]: INFO nova.compute.manager [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Took 13.67 seconds to build instance. [ 1435.829116] env[65385]: DEBUG nova.compute.manager [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Received event network-changed-f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1435.829324] env[65385]: DEBUG nova.compute.manager [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Refreshing instance network info cache due to event network-changed-f60c60da-40d3-44a8-b2bd-0d0464f0e773. {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1435.829517] env[65385]: DEBUG oslo_concurrency.lockutils [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Acquiring lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1435.829663] env[65385]: DEBUG oslo_concurrency.lockutils [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Acquired lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1435.829822] env[65385]: DEBUG nova.network.neutron [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Refreshing network info cache for port f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1435.908545] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: b92bc73a-9666-435e-b387-18df7c838f9b] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1436.050339] env[65385]: DEBUG oslo_concurrency.lockutils [None req-daf9c53a-b087-43ee-bff9-dd803a4f3192 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.178s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1436.332283] env[65385]: WARNING neutronclient.v2_0.client [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1436.333039] env[65385]: WARNING openstack [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1436.333397] env[65385]: WARNING openstack [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1436.412646] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 7f22736b-654f-4d51-82d4-3fe80c1d0b52] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1436.443503] env[65385]: WARNING openstack [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1436.443957] env[65385]: WARNING openstack [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1436.505828] env[65385]: WARNING neutronclient.v2_0.client [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1436.506512] env[65385]: WARNING openstack [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1436.507076] env[65385]: WARNING openstack [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1436.587464] env[65385]: DEBUG nova.network.neutron [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updated VIF entry in instance network info cache for port f60c60da-40d3-44a8-b2bd-0d0464f0e773. {{(pid=65385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1436.587886] env[65385]: DEBUG nova.network.neutron [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updating instance_info_cache with network_info: [{"id": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "address": "fa:16:3e:5a:f9:f5", "network": {"id": "77ff37c4-51ac-428b-ab8e-ab204ba35f5c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-519318608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df347ee234f142899014703c70ee42ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ebd8af-aaf6-4d04-b869-3882e2571ed7", "external-id": "nsx-vlan-transportzone-541", "segmentation_id": 541, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf60c60da-40", "ovs_interfaceid": "f60c60da-40d3-44a8-b2bd-0d0464f0e773", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1436.915656] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: df419705-e7e8-47b4-b9a0-2f1cf2638f33] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1437.090541] env[65385]: DEBUG oslo_concurrency.lockutils [req-095179aa-db27-41f1-b5a2-dcdf0d25fe94 req-549a5142-bd29-406b-abc9-6193557d54e6 service nova] Releasing lock "refresh_cache-218379e3-49ca-4fdf-8c21-757ae53a1653" {{(pid=65385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1437.419308] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: b5f563d1-f0e6-4521-8e07-7bbdc5d42a4f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1437.922858] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 805528b8-96db-4bb8-ba05-2b6332c54624] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1438.426242] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 6c826914-4c6b-456b-a403-2373d30c803f] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1438.930404] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 900639b6-9b98-436f-aaad-ea8391f67393] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1439.434225] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: b1abf6ab-a3d5-4078-9568-68db0a7f03cc] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1439.938475] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 992920c3-af4f-489e-b21f-9d52c24399de] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1440.442319] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: b4bca2c8-fc8f-471a-899b-aac648fee9fe] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1440.945608] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] [instance: 29a449d0-f62d-4430-96ab-9ac43e5a7ad2] Instance has had 0 of 5 cleanup attempts {{(pid=65385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1441.449921] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1442.951869] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1467.220503] env[65385]: DEBUG oslo_concurrency.lockutils [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1467.221208] env[65385]: DEBUG oslo_concurrency.lockutils [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1467.593968] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1467.594262] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1467.724821] env[65385]: DEBUG nova.compute.utils [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1468.097512] env[65385]: INFO nova.compute.manager [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Detaching volume 4d9484bb-eb7b-4915-ab78-a5d252589350 [ 1468.127533] env[65385]: INFO nova.virt.block_device [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Attempting to driver detach volume 4d9484bb-eb7b-4915-ab78-a5d252589350 from mountpoint /dev/sdb [ 1468.127775] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1468.127955] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871225', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'name': 'volume-4d9484bb-eb7b-4915-ab78-a5d252589350', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9ee16fed-c625-4da3-bb83-2b4b2416b99f', 'attached_at': '', 'detached_at': '', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'serial': '4d9484bb-eb7b-4915-ab78-a5d252589350'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1468.128907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279f68dd-6eb6-4f66-9584-9d47eb0aee09 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.151233] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c1122c-a57b-451b-a02b-82243bdc807f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.158586] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30398e15-8c96-4e9d-a324-5f058f065a7a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.178907] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d006b7-fcb4-44ae-a2f0-e0810ad98dfc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.194626] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] The volume has not been displaced from its original location: [datastore1] volume-4d9484bb-eb7b-4915-ab78-a5d252589350/volume-4d9484bb-eb7b-4915-ab78-a5d252589350.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1468.199743] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfiguring VM instance instance-0000007e to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1468.200067] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b371d868-1c07-44df-87cf-64f2887b6d47 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.218345] env[65385]: DEBUG oslo_vmware.api [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1468.218345] env[65385]: value = "task-4454838" [ 1468.218345] env[65385]: _type = "Task" [ 1468.218345] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1468.226427] env[65385]: DEBUG oslo_vmware.api [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454838, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1468.226972] env[65385]: DEBUG oslo_concurrency.lockutils [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1468.729513] env[65385]: DEBUG oslo_vmware.api [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454838, 'name': ReconfigVM_Task, 'duration_secs': 0.244542} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1468.729513] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Reconfigured VM instance instance-0000007e to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1468.734182] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42928d54-ffa7-49f1-a207-f9381a340b56 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.750402] env[65385]: DEBUG oslo_vmware.api [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1468.750402] env[65385]: value = "task-4454839" [ 1468.750402] env[65385]: _type = "Task" [ 1468.750402] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1468.759165] env[65385]: DEBUG oslo_vmware.api [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454839, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1469.261226] env[65385]: DEBUG oslo_vmware.api [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454839, 'name': ReconfigVM_Task, 'duration_secs': 0.143548} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1469.261609] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871225', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'name': 'volume-4d9484bb-eb7b-4915-ab78-a5d252589350', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9ee16fed-c625-4da3-bb83-2b4b2416b99f', 'attached_at': '', 'detached_at': '', 'volume_id': '4d9484bb-eb7b-4915-ab78-a5d252589350', 'serial': '4d9484bb-eb7b-4915-ab78-a5d252589350'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1469.290123] env[65385]: DEBUG oslo_concurrency.lockutils [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1469.290389] env[65385]: DEBUG oslo_concurrency.lockutils [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1469.290620] env[65385]: INFO nova.compute.manager [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Attaching volume cbf73fcd-a3b8-4c48-bff6-5b2267900223 to /dev/sdb [ 1469.320117] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e392cc88-9dd9-4585-b678-c7924db875e4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.328147] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3788a92-f021-4a80-bfe4-5827e73ec540 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.341525] env[65385]: DEBUG nova.virt.block_device [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updating existing volume attachment record: dbd2e7e5-e6cd-41af-a70d-2d421c5e7526 {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1469.805693] env[65385]: DEBUG nova.objects.instance [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lazy-loading 'flavor' on Instance uuid 9ee16fed-c625-4da3-bb83-2b4b2416b99f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1470.815156] env[65385]: DEBUG oslo_concurrency.lockutils [None req-fdb00be2-51ac-4e76-abf3-081d9fd58671 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.221s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1471.904156] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1471.904466] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1471.904619] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1471.904798] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1471.904964] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1471.907153] env[65385]: INFO nova.compute.manager [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Terminating instance [ 1472.410979] env[65385]: DEBUG nova.compute.manager [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1472.411227] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1472.412136] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699aaaf4-9683-41b4-a16e-8d77eb2819db {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1472.420525] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1472.420766] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e38681dd-caf1-4bf4-9e89-a41af4990999 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1472.427067] env[65385]: DEBUG oslo_vmware.api [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1472.427067] env[65385]: value = "task-4454842" [ 1472.427067] env[65385]: _type = "Task" [ 1472.427067] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1472.435920] env[65385]: DEBUG oslo_vmware.api [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1472.937096] env[65385]: DEBUG oslo_vmware.api [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454842, 'name': PowerOffVM_Task, 'duration_secs': 0.1868} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1472.937410] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1472.937561] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1472.937825] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7e65461-d28f-41df-b662-ff018720942f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.006319] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1473.006523] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1473.006688] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleting the datastore file [datastore2] 9ee16fed-c625-4da3-bb83-2b4b2416b99f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1473.006932] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f38fb6d7-1414-4b51-893b-91558d5668e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.013801] env[65385]: DEBUG oslo_vmware.api [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1473.013801] env[65385]: value = "task-4454844" [ 1473.013801] env[65385]: _type = "Task" [ 1473.013801] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1473.022059] env[65385]: DEBUG oslo_vmware.api [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1473.524593] env[65385]: DEBUG oslo_vmware.api [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156712} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1473.524848] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1473.525033] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1473.525206] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1473.525406] env[65385]: INFO nova.compute.manager [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1473.525655] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1473.525848] env[65385]: DEBUG nova.compute.manager [-] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1473.525941] env[65385]: DEBUG nova.network.neutron [-] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1473.526213] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1473.526737] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1473.526988] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1473.563602] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1473.885451] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1473.885699] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871227', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'name': 'volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'serial': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1473.886626] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb5b4f0-58ba-43e6-92a2-79f683b911e1 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.903203] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da6788d-e153-48fa-aaf0-0c963fa1394f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.927585] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfiguring VM instance instance-0000007f to attach disk [datastore1] volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223/volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1473.927883] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ca875e2-b927-4279-85cb-21d7037179b2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1473.947618] env[65385]: DEBUG oslo_vmware.api [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1473.947618] env[65385]: value = "task-4454845" [ 1473.947618] env[65385]: _type = "Task" [ 1473.947618] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1473.956263] env[65385]: DEBUG oslo_vmware.api [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454845, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1474.030013] env[65385]: DEBUG nova.compute.manager [req-850121ac-a48c-4329-a123-4e05f5d09691 req-700a111e-9951-4680-aa10-923e340de3b3 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Received event network-vif-deleted-c6668bfd-25c6-4589-b8ee-bbbc79181617 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1474.030220] env[65385]: INFO nova.compute.manager [req-850121ac-a48c-4329-a123-4e05f5d09691 req-700a111e-9951-4680-aa10-923e340de3b3 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Neutron deleted interface c6668bfd-25c6-4589-b8ee-bbbc79181617; detaching it from the instance and deleting it from the info cache [ 1474.030581] env[65385]: DEBUG nova.network.neutron [req-850121ac-a48c-4329-a123-4e05f5d09691 req-700a111e-9951-4680-aa10-923e340de3b3 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1474.458338] env[65385]: DEBUG oslo_vmware.api [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454845, 'name': ReconfigVM_Task, 'duration_secs': 0.33874} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1474.458671] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfigured VM instance instance-0000007f to attach disk [datastore1] volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223/volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1474.463255] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c962ef4-ed32-47fb-a9b2-75a20b98068a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1474.479950] env[65385]: DEBUG oslo_vmware.api [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1474.479950] env[65385]: value = "task-4454846" [ 1474.479950] env[65385]: _type = "Task" [ 1474.479950] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1474.489357] env[65385]: DEBUG oslo_vmware.api [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1474.501889] env[65385]: DEBUG nova.network.neutron [-] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1474.532851] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-57f4e2a0-17e6-45b8-bcf2-70fac6fe83da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1474.543276] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200cfe76-3c65-4ab2-a9b0-df8d3ee16cf4 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1474.571938] env[65385]: DEBUG nova.compute.manager [req-850121ac-a48c-4329-a123-4e05f5d09691 req-700a111e-9951-4680-aa10-923e340de3b3 service nova] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Detach interface failed, port_id=c6668bfd-25c6-4589-b8ee-bbbc79181617, reason: Instance 9ee16fed-c625-4da3-bb83-2b4b2416b99f could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1474.990486] env[65385]: DEBUG oslo_vmware.api [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454846, 'name': ReconfigVM_Task, 'duration_secs': 0.175503} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1474.990843] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871227', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'name': 'volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'serial': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1475.005488] env[65385]: INFO nova.compute.manager [-] [instance: 9ee16fed-c625-4da3-bb83-2b4b2416b99f] Took 1.48 seconds to deallocate network for instance. [ 1475.513149] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1475.513457] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1475.513682] env[65385]: DEBUG nova.objects.instance [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lazy-loading 'resources' on Instance uuid 9ee16fed-c625-4da3-bb83-2b4b2416b99f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1476.029451] env[65385]: DEBUG nova.objects.instance [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 218379e3-49ca-4fdf-8c21-757ae53a1653 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1476.087234] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd10b81d-7b44-49ae-a94a-ca6ad70fbb0a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.096025] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52efdf7f-6dc1-4f0a-bafb-ee9a0ba6c87a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.127988] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92944099-797c-424f-932c-9dfcbc2efa01 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.136103] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45632c11-e930-4e24-915b-e0a773af9c35 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.150213] env[65385]: DEBUG nova.compute.provider_tree [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1476.537204] env[65385]: DEBUG oslo_concurrency.lockutils [None req-86f60b2c-e8c9-416c-bac8-973cdfedaff1 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1476.653406] env[65385]: DEBUG nova.scheduler.client.report [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1477.159078] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1477.183273] env[65385]: INFO nova.scheduler.client.report [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleted allocations for instance 9ee16fed-c625-4da3-bb83-2b4b2416b99f [ 1477.386091] env[65385]: DEBUG oslo_concurrency.lockutils [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1477.386330] env[65385]: DEBUG oslo_concurrency.lockutils [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1477.690974] env[65385]: DEBUG oslo_concurrency.lockutils [None req-26d0726f-55a3-4358-b7aa-50813ab290de tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "9ee16fed-c625-4da3-bb83-2b4b2416b99f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.786s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1477.890107] env[65385]: DEBUG nova.compute.utils [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Using /dev/sd instead of None {{(pid=65385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1478.393822] env[65385]: DEBUG oslo_concurrency.lockutils [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1479.456595] env[65385]: DEBUG oslo_concurrency.lockutils [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1479.456995] env[65385]: DEBUG oslo_concurrency.lockutils [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1479.457136] env[65385]: INFO nova.compute.manager [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Attaching volume ce61f0b0-7a8d-4686-9de7-f46fd170273a to /dev/sdc [ 1479.515302] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993e3abb-ef11-4efe-b32d-3f0e0f01a9d6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.523090] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4c7cea-207b-4c41-b638-14516e52116a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.536858] env[65385]: DEBUG nova.virt.block_device [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updating existing volume attachment record: 17b10034-2394-4701-a216-8d4a890e5bfc {{(pid=65385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1481.844318] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1481.844811] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager.update_available_resource {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1482.347495] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1482.347762] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1482.347930] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1482.348096] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1482.349017] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f63587f-9ca0-40ff-98e8-2ac43d46850c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.357904] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04b8fec-bf88-4528-a3dd-ed9111405557 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.372149] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1dd1a94-d9af-45fa-a4ce-11dac1752884 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.378934] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3ec7db-ff54-4007-a3c6-7bec1ced6d0f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.409153] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179729MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1482.409367] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1482.409662] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1483.185610] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1483.185970] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1483.186075] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1483.186289] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1483.186464] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1483.190267] env[65385]: INFO nova.compute.manager [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Terminating instance [ 1483.436166] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 02e65e47-7971-4e08-acc9-eead1828d925 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1483.436328] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance ad85aa5c-68a6-41a3-88d7-c27bbef81c6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1483.436447] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Instance 218379e3-49ca-4fdf-8c21-757ae53a1653 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1483.436620] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1483.436764] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=100GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '3', 'num_vm_active': '2', 'num_task_None': '3', 'num_os_type_None': '3', 'num_proj_2fc382ad43724fd3ae1b9b62c1dd8131': '2', 'io_workload': '0', 'num_vm_rescued': '1', 'num_proj_df347ee234f142899014703c70ee42ed': '1'} {{(pid=65385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1483.488583] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97126631-2394-42f5-bb25-9de6b0f03f9a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.496697] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ca86b8-8a7d-438d-88e1-8ed9529bf854 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.526811] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179cfc6f-10ad-4572-8ac5-8298a5f133da {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.534962] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606912ea-68ba-4bcb-b6b1-934012e09cab {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.549604] env[65385]: DEBUG nova.compute.provider_tree [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1483.694155] env[65385]: DEBUG nova.compute.manager [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1483.694336] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1483.695267] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b75d8b-a24b-4c25-baa9-d4a8ee6af486 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.703680] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1483.703909] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14c2ec13-2adf-4345-bb3e-0957cef71bee {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.709750] env[65385]: DEBUG oslo_vmware.api [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1483.709750] env[65385]: value = "task-4454850" [ 1483.709750] env[65385]: _type = "Task" [ 1483.709750] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1483.718337] env[65385]: DEBUG oslo_vmware.api [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454850, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.052814] env[65385]: DEBUG nova.scheduler.client.report [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1484.081716] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Volume attach. Driver type: vmdk {{(pid=65385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1484.081958] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871228', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'name': 'volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'serial': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1484.082837] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e757b8-ce36-4bbd-8aa8-1d2af77ee0dd {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.099965] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39fdeb52-f488-46fe-94dd-f539b770b220 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.129839] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfiguring VM instance instance-0000007f to attach disk [datastore2] volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a/volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1484.130157] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53c4f6e2-c789-4b5d-a3a6-25f88ec15c5d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.149342] env[65385]: DEBUG oslo_vmware.api [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1484.149342] env[65385]: value = "task-4454851" [ 1484.149342] env[65385]: _type = "Task" [ 1484.149342] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.158204] env[65385]: DEBUG oslo_vmware.api [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454851, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.219710] env[65385]: DEBUG oslo_vmware.api [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454850, 'name': PowerOffVM_Task, 'duration_secs': 0.194246} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.220094] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1484.220267] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1484.220523] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a0849ef-41ce-4a4f-b1e7-8f2330c77b45 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.286216] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1484.286470] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1484.286674] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleting the datastore file [datastore2] ad85aa5c-68a6-41a3-88d7-c27bbef81c6f {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1484.286984] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99223ffd-8259-4d4f-939f-eb8184a491be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.295838] env[65385]: DEBUG oslo_vmware.api [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1484.295838] env[65385]: value = "task-4454853" [ 1484.295838] env[65385]: _type = "Task" [ 1484.295838] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.304812] env[65385]: DEBUG oslo_vmware.api [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454853, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.557655] env[65385]: DEBUG nova.compute.resource_tracker [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1484.557846] env[65385]: DEBUG oslo_concurrency.lockutils [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.148s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1484.660607] env[65385]: DEBUG oslo_vmware.api [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454851, 'name': ReconfigVM_Task, 'duration_secs': 0.376638} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.660975] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfigured VM instance instance-0000007f to attach disk [datastore2] volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a/volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a.vmdk or device None with type thin {{(pid=65385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1484.666362] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5468307-4cae-4456-ab6f-2e2928425c02 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.683914] env[65385]: DEBUG oslo_vmware.api [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1484.683914] env[65385]: value = "task-4454854" [ 1484.683914] env[65385]: _type = "Task" [ 1484.683914] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.695760] env[65385]: DEBUG oslo_vmware.api [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454854, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.806015] env[65385]: DEBUG oslo_vmware.api [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454853, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194745} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.806270] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1484.806442] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1484.806614] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1484.806782] env[65385]: INFO nova.compute.manager [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1484.807015] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1484.807213] env[65385]: DEBUG nova.compute.manager [-] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1484.807311] env[65385]: DEBUG nova.network.neutron [-] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1484.807552] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1484.808128] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1484.808361] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1484.854273] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1485.119043] env[65385]: DEBUG nova.compute.manager [req-331122bf-692c-4d99-b38f-5e67f70286a4 req-b3c4d19c-7a94-48c4-b7d5-3231ee9b0206 service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Received event network-vif-deleted-764d3f51-03a5-48ff-bab9-f005c7613395 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1485.119269] env[65385]: INFO nova.compute.manager [req-331122bf-692c-4d99-b38f-5e67f70286a4 req-b3c4d19c-7a94-48c4-b7d5-3231ee9b0206 service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Neutron deleted interface 764d3f51-03a5-48ff-bab9-f005c7613395; detaching it from the instance and deleting it from the info cache [ 1485.119424] env[65385]: DEBUG nova.network.neutron [req-331122bf-692c-4d99-b38f-5e67f70286a4 req-b3c4d19c-7a94-48c4-b7d5-3231ee9b0206 service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1485.195690] env[65385]: DEBUG oslo_vmware.api [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454854, 'name': ReconfigVM_Task, 'duration_secs': 0.255987} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1485.196097] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871228', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'name': 'volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'serial': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a'} {{(pid=65385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1485.605318] env[65385]: DEBUG nova.network.neutron [-] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1485.622286] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7883f8e2-c186-4606-baae-813d2bd2c492 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.635460] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff527621-f205-4ccc-8cec-9c07c8fe1a6d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.664494] env[65385]: DEBUG nova.compute.manager [req-331122bf-692c-4d99-b38f-5e67f70286a4 req-b3c4d19c-7a94-48c4-b7d5-3231ee9b0206 service nova] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Detach interface failed, port_id=764d3f51-03a5-48ff-bab9-f005c7613395, reason: Instance ad85aa5c-68a6-41a3-88d7-c27bbef81c6f could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1486.108188] env[65385]: INFO nova.compute.manager [-] [instance: ad85aa5c-68a6-41a3-88d7-c27bbef81c6f] Took 1.30 seconds to deallocate network for instance. [ 1486.241769] env[65385]: DEBUG nova.objects.instance [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 218379e3-49ca-4fdf-8c21-757ae53a1653 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1486.551522] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1486.551734] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1486.614475] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1486.614884] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1486.614969] env[65385]: DEBUG nova.objects.instance [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lazy-loading 'resources' on Instance uuid ad85aa5c-68a6-41a3-88d7-c27bbef81c6f {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1486.747135] env[65385]: DEBUG oslo_concurrency.lockutils [None req-24203576-27a3-4b50-8319-905fc9ba0a2b tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.290s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1487.042029] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1487.042242] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1487.135069] env[65385]: DEBUG nova.scheduler.client.report [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Refreshing inventories for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1487.149041] env[65385]: DEBUG nova.scheduler.client.report [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Updating ProviderTree inventory for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1487.149333] env[65385]: DEBUG nova.compute.provider_tree [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Updating inventory in ProviderTree for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1487.160831] env[65385]: DEBUG nova.scheduler.client.report [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Refreshing aggregate associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, aggregates: None {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1487.178905] env[65385]: DEBUG nova.scheduler.client.report [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Refreshing trait associations for resource provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1487.229616] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07435a9b-c975-4b1b-b9fe-778d821b426c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.237721] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf15d028-1c53-4d41-8222-3769076ca308 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.268100] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779fb491-9ca7-4ce6-b6bc-f589af3acae9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.276473] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0f60f5-a637-4bbd-8dea-c55ba046f5f8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.290523] env[65385]: DEBUG nova.compute.provider_tree [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1487.545406] env[65385]: INFO nova.compute.manager [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Detaching volume cbf73fcd-a3b8-4c48-bff6-5b2267900223 [ 1487.581452] env[65385]: INFO nova.virt.block_device [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Attempting to driver detach volume cbf73fcd-a3b8-4c48-bff6-5b2267900223 from mountpoint /dev/sdb [ 1487.581697] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1487.581891] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871227', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'name': 'volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'serial': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1487.582851] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8836a5-9afd-4b42-bd73-edfe32a65d3b {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.609474] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33843f64-a8ba-45b6-903b-474ce5c41e3c {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.617179] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d64bf37-0682-4926-8299-79d03c5b6b52 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.641732] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7ad588-3c1e-48fd-a8fc-a288dc18ea0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.657906] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The volume has not been displaced from its original location: [datastore1] volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223/volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1487.663359] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfiguring VM instance instance-0000007f to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1487.663695] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58433e0f-2dd0-4271-bd62-315439dd94b0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.683023] env[65385]: DEBUG oslo_vmware.api [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1487.683023] env[65385]: value = "task-4454855" [ 1487.683023] env[65385]: _type = "Task" [ 1487.683023] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1487.692291] env[65385]: DEBUG oslo_vmware.api [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1487.794389] env[65385]: DEBUG nova.scheduler.client.report [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1487.844277] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1487.844480] env[65385]: DEBUG nova.compute.manager [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1488.193476] env[65385]: DEBUG oslo_vmware.api [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454855, 'name': ReconfigVM_Task, 'duration_secs': 0.224774} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1488.193748] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfigured VM instance instance-0000007f to detach disk 2001 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1488.198339] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b435ef0-40cc-4380-b159-7e93941d382e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1488.214078] env[65385]: DEBUG oslo_vmware.api [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1488.214078] env[65385]: value = "task-4454856" [ 1488.214078] env[65385]: _type = "Task" [ 1488.214078] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1488.226063] env[65385]: DEBUG oslo_vmware.api [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1488.301434] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1488.333465] env[65385]: INFO nova.scheduler.client.report [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleted allocations for instance ad85aa5c-68a6-41a3-88d7-c27bbef81c6f [ 1488.725689] env[65385]: DEBUG oslo_vmware.api [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454856, 'name': ReconfigVM_Task, 'duration_secs': 0.149011} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1488.726101] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871227', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'name': 'volume-cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223', 'serial': 'cbf73fcd-a3b8-4c48-bff6-5b2267900223'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1488.843144] env[65385]: DEBUG oslo_concurrency.lockutils [None req-aa9b0300-3a04-4198-bf40-69d540c811c9 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "ad85aa5c-68a6-41a3-88d7-c27bbef81c6f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.657s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1488.844378] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1488.844492] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1489.268255] env[65385]: DEBUG nova.objects.instance [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 218379e3-49ca-4fdf-8c21-757ae53a1653 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1489.843513] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1489.927651] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "02e65e47-7971-4e08-acc9-eead1828d925" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1489.927967] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "02e65e47-7971-4e08-acc9-eead1828d925" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1489.928203] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "02e65e47-7971-4e08-acc9-eead1828d925-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1489.928384] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "02e65e47-7971-4e08-acc9-eead1828d925-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1489.928553] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "02e65e47-7971-4e08-acc9-eead1828d925-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1489.930781] env[65385]: INFO nova.compute.manager [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Terminating instance [ 1490.276214] env[65385]: DEBUG oslo_concurrency.lockutils [None req-ebd2b166-0468-4b80-a45d-320e3a4cf589 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.234s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1490.316690] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1490.316938] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1490.434319] env[65385]: DEBUG nova.compute.manager [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1490.434546] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1490.435470] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68030dca-a654-4607-8763-9efae257e4be {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.443948] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1490.444201] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-365879ed-834f-4bdf-a2f9-35ce1086b8f9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.451366] env[65385]: DEBUG oslo_vmware.api [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1490.451366] env[65385]: value = "task-4454857" [ 1490.451366] env[65385]: _type = "Task" [ 1490.451366] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.460837] env[65385]: DEBUG oslo_vmware.api [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454857, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.820563] env[65385]: INFO nova.compute.manager [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Detaching volume ce61f0b0-7a8d-4686-9de7-f46fd170273a [ 1490.858653] env[65385]: INFO nova.virt.block_device [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Attempting to driver detach volume ce61f0b0-7a8d-4686-9de7-f46fd170273a from mountpoint /dev/sdc [ 1490.858945] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Volume detach. Driver type: vmdk {{(pid=65385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1490.859153] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871228', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'name': 'volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'serial': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1490.860047] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533441c4-e76e-47ad-9379-b1662e706b73 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.883841] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797420b2-fe0b-4b94-a249-59f2e61f7f0e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.892128] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51277d18-3f7f-4dbc-b525-516c2ba52ecf {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.913231] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a9c2f9-2938-47c9-b2d4-ecc3a896fec8 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.928848] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] The volume has not been displaced from its original location: [datastore2] volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a/volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a.vmdk. No consolidation needed. {{(pid=65385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1490.933952] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfiguring VM instance instance-0000007f to detach disk 2002 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1490.934268] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-604dea01-4a56-4b80-8917-ed1e59498b83 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.956455] env[65385]: DEBUG oslo_vmware.api [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1490.956455] env[65385]: value = "task-4454858" [ 1490.956455] env[65385]: _type = "Task" [ 1490.956455] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.962726] env[65385]: DEBUG oslo_vmware.api [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454857, 'name': PowerOffVM_Task, 'duration_secs': 0.199557} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1490.963360] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1490.963534] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1490.963810] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97ff1307-4d4a-43a6-8be9-2d9bb37bb79e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.968526] env[65385]: DEBUG oslo_vmware.api [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454858, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.032367] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1491.032650] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Deleting contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1491.032855] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleting the datastore file [datastore2] 02e65e47-7971-4e08-acc9-eead1828d925 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1491.033157] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d8a1b14-7ab4-4d94-aa68-fd7943d0804e {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.039557] env[65385]: DEBUG oslo_vmware.api [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for the task: (returnval){ [ 1491.039557] env[65385]: value = "task-4454860" [ 1491.039557] env[65385]: _type = "Task" [ 1491.039557] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1491.058783] env[65385]: DEBUG oslo_vmware.api [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454860, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.466507] env[65385]: DEBUG oslo_vmware.api [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454858, 'name': ReconfigVM_Task, 'duration_secs': 0.241201} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.466847] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Reconfigured VM instance instance-0000007f to detach disk 2002 {{(pid=65385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1491.471398] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6231a536-b58a-4985-bb1d-46bf73defdfc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.486619] env[65385]: DEBUG oslo_vmware.api [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1491.486619] env[65385]: value = "task-4454861" [ 1491.486619] env[65385]: _type = "Task" [ 1491.486619] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1491.495540] env[65385]: DEBUG oslo_vmware.api [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454861, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.550584] env[65385]: DEBUG oslo_vmware.api [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Task: {'id': task-4454860, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176024} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.550831] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1491.551023] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Deleted contents of the VM from datastore datastore2 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1491.551204] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1491.551374] env[65385]: INFO nova.compute.manager [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1491.551609] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1491.551805] env[65385]: DEBUG nova.compute.manager [-] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1491.551903] env[65385]: DEBUG nova.network.neutron [-] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1491.552169] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1491.552693] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1491.552950] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1491.605473] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1491.839280] env[65385]: DEBUG oslo_service.periodic_task [None req-8dbf0710-3af4-4452-bdd2-786d32ec0f7f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1491.892173] env[65385]: DEBUG nova.compute.manager [req-a5c08b25-bda0-42ee-8912-ce33f3322856 req-d6744cc5-f755-4189-9ef4-febaf4b5a8f4 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Received event network-vif-deleted-078edaec-5c64-4393-b4b9-0e9537ee060c {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1491.892448] env[65385]: INFO nova.compute.manager [req-a5c08b25-bda0-42ee-8912-ce33f3322856 req-d6744cc5-f755-4189-9ef4-febaf4b5a8f4 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Neutron deleted interface 078edaec-5c64-4393-b4b9-0e9537ee060c; detaching it from the instance and deleting it from the info cache [ 1491.892542] env[65385]: DEBUG nova.network.neutron [req-a5c08b25-bda0-42ee-8912-ce33f3322856 req-d6744cc5-f755-4189-9ef4-febaf4b5a8f4 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1492.000746] env[65385]: DEBUG oslo_vmware.api [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454861, 'name': ReconfigVM_Task, 'duration_secs': 0.138052} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1492.001071] env[65385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-871228', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'name': 'volume-ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '218379e3-49ca-4fdf-8c21-757ae53a1653', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a', 'serial': 'ce61f0b0-7a8d-4686-9de7-f46fd170273a'} {{(pid=65385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1492.365955] env[65385]: DEBUG nova.network.neutron [-] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1492.394891] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8b3b20a-3880-42e8-834e-88387f528160 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.405499] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073a6d62-423a-4504-ba60-865dbd706869 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.431774] env[65385]: DEBUG nova.compute.manager [req-a5c08b25-bda0-42ee-8912-ce33f3322856 req-d6744cc5-f755-4189-9ef4-febaf4b5a8f4 service nova] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Detach interface failed, port_id=078edaec-5c64-4393-b4b9-0e9537ee060c, reason: Instance 02e65e47-7971-4e08-acc9-eead1828d925 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1492.545128] env[65385]: DEBUG nova.objects.instance [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'flavor' on Instance uuid 218379e3-49ca-4fdf-8c21-757ae53a1653 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1492.868400] env[65385]: INFO nova.compute.manager [-] [instance: 02e65e47-7971-4e08-acc9-eead1828d925] Took 1.32 seconds to deallocate network for instance. [ 1493.376152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1493.376152] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1493.376580] env[65385]: DEBUG nova.objects.instance [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lazy-loading 'resources' on Instance uuid 02e65e47-7971-4e08-acc9-eead1828d925 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1493.552567] env[65385]: DEBUG oslo_concurrency.lockutils [None req-b74856f1-c085-48b6-820b-a5ffa8d856bc tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.235s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1493.926680] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d92151c-fc63-4a4c-92ad-071932d1033f {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.935357] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d475e53c-5581-4aad-9921-cd9c27ca62b6 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.970498] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9b4cb3-4eb8-4172-95ce-8f2e4807b8bb {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.979581] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc44874-461f-4831-b65d-187e8fd6fa1d {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.993852] env[65385]: DEBUG nova.compute.provider_tree [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1494.497378] env[65385]: DEBUG nova.scheduler.client.report [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1494.715501] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1494.715847] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1494.716133] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "218379e3-49ca-4fdf-8c21-757ae53a1653-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1494.716370] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1494.716616] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1494.719310] env[65385]: INFO nova.compute.manager [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Terminating instance [ 1495.003104] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.627s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1495.025474] env[65385]: INFO nova.scheduler.client.report [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Deleted allocations for instance 02e65e47-7971-4e08-acc9-eead1828d925 [ 1495.223596] env[65385]: DEBUG nova.compute.manager [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Start destroying the instance on the hypervisor. {{(pid=65385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1495.223596] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Destroying instance {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1495.224504] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27149508-91e3-4c10-994c-265e4f385885 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.233376] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Powering off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1495.233689] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d45f6002-4bc5-4203-82b1-d9b929dada2a {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.241289] env[65385]: DEBUG oslo_vmware.api [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1495.241289] env[65385]: value = "task-4454862" [ 1495.241289] env[65385]: _type = "Task" [ 1495.241289] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1495.253564] env[65385]: DEBUG oslo_vmware.api [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454862, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1495.533702] env[65385]: DEBUG oslo_concurrency.lockutils [None req-e5ec29bd-9c7f-448b-9343-b68f02c74541 tempest-ServerRescueNegativeTestJSON-1614036554 tempest-ServerRescueNegativeTestJSON-1614036554-project-member] Lock "02e65e47-7971-4e08-acc9-eead1828d925" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.606s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1495.751258] env[65385]: DEBUG oslo_vmware.api [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454862, 'name': PowerOffVM_Task, 'duration_secs': 0.216377} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1495.751541] env[65385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Powered off the VM {{(pid=65385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1495.751724] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Unregistering the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1495.751987] env[65385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dd5d868-e2b9-47fc-bad3-2a36dadeeee2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.819972] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Unregistered the VM {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1495.820363] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Deleting contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1495.820660] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Deleting the datastore file [datastore1] 218379e3-49ca-4fdf-8c21-757ae53a1653 {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1495.821057] env[65385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b792815-0d17-423c-ad1c-26443a4f83dc {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.833762] env[65385]: DEBUG oslo_vmware.api [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for the task: (returnval){ [ 1495.833762] env[65385]: value = "task-4454864" [ 1495.833762] env[65385]: _type = "Task" [ 1495.833762] env[65385]: } to complete. {{(pid=65385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1495.842442] env[65385]: DEBUG oslo_vmware.api [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454864, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1496.345812] env[65385]: DEBUG oslo_vmware.api [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Task: {'id': task-4454864, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146181} completed successfully. {{(pid=65385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1496.345812] env[65385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Deleted the datastore file {{(pid=65385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1496.345812] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Deleted contents of the VM from datastore datastore1 {{(pid=65385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1496.345812] env[65385]: DEBUG nova.virt.vmwareapi.vmops [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Instance destroyed {{(pid=65385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1496.345812] env[65385]: INFO nova.compute.manager [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1496.345812] env[65385]: DEBUG oslo.service.backend._eventlet.loopingcall [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/loopingcall.py:437}} [ 1496.345812] env[65385]: DEBUG nova.compute.manager [-] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Deallocating network for instance {{(pid=65385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1496.346543] env[65385]: DEBUG nova.network.neutron [-] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] deallocate_for_instance() {{(pid=65385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1496.346543] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1496.346663] env[65385]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1496.346919] env[65385]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1496.382415] env[65385]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1496.820092] env[65385]: DEBUG nova.compute.manager [req-7ed3dab3-a172-4a9b-a951-b095d5f30b31 req-ba51fec7-b558-4d9d-bcaa-d36c6f157b25 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Received event network-vif-deleted-f60c60da-40d3-44a8-b2bd-0d0464f0e773 {{(pid=65385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1496.820492] env[65385]: INFO nova.compute.manager [req-7ed3dab3-a172-4a9b-a951-b095d5f30b31 req-ba51fec7-b558-4d9d-bcaa-d36c6f157b25 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Neutron deleted interface f60c60da-40d3-44a8-b2bd-0d0464f0e773; detaching it from the instance and deleting it from the info cache [ 1496.820639] env[65385]: DEBUG nova.network.neutron [req-7ed3dab3-a172-4a9b-a951-b095d5f30b31 req-ba51fec7-b558-4d9d-bcaa-d36c6f157b25 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1497.284401] env[65385]: DEBUG nova.network.neutron [-] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Updating instance_info_cache with network_info: [] {{(pid=65385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1497.323206] env[65385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e411e36-79aa-4768-9060-4c6fe471aee0 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.334141] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949b7366-f97f-435c-9712-e28a79af3d60 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.361157] env[65385]: DEBUG nova.compute.manager [req-7ed3dab3-a172-4a9b-a951-b095d5f30b31 req-ba51fec7-b558-4d9d-bcaa-d36c6f157b25 service nova] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Detach interface failed, port_id=f60c60da-40d3-44a8-b2bd-0d0464f0e773, reason: Instance 218379e3-49ca-4fdf-8c21-757ae53a1653 could not be found. {{(pid=65385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1497.788285] env[65385]: INFO nova.compute.manager [-] [instance: 218379e3-49ca-4fdf-8c21-757ae53a1653] Took 1.44 seconds to deallocate network for instance. [ 1498.298973] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1498.299355] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1498.299511] env[65385]: DEBUG nova.objects.instance [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lazy-loading 'resources' on Instance uuid 218379e3-49ca-4fdf-8c21-757ae53a1653 {{(pid=65385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1498.835770] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789ae0e3-ba45-4544-a6b9-5858fd7923a2 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.843899] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a937643-67b0-4da2-a66b-3744470ef919 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.878430] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c769d023-f6b0-41e2-879c-8145b83c6252 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.886666] env[65385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cb222a-204e-4e33-8b0b-cf8afd649eb9 {{(pid=65385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1498.900928] env[65385]: DEBUG nova.compute.provider_tree [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed in ProviderTree for provider: 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b {{(pid=65385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1499.404767] env[65385]: DEBUG nova.scheduler.client.report [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Inventory has not changed for provider 1af23b69-5ce6-4d6c-8591-1b95ecca8a6b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1499.909641] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1499.929392] env[65385]: INFO nova.scheduler.client.report [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Deleted allocations for instance 218379e3-49ca-4fdf-8c21-757ae53a1653 [ 1500.439785] env[65385]: DEBUG oslo_concurrency.lockutils [None req-a529ed32-6ac5-44bc-8adb-9e2d3dee9057 tempest-AttachVolumeTestJSON-1147150116 tempest-AttachVolumeTestJSON-1147150116-project-member] Lock "218379e3-49ca-4fdf-8c21-757ae53a1653" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.723s {{(pid=65385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}}